![]() | Computer Graphics -- Image-based Modeling and Rendering, View Synthesis |
![]() | Computer Vision -- 3D Modeling, Real-time Vision (stereo, tracking, etc) |
![]() | Multimedia -- Video Teleconferencing, Automatic Meeting Management, HCI |
![]() | Image Processing -- Real-time image processing |
![]() | Graphics Hardware -- Novel Applications on Graphics Hardware |
3D Physically-based 2D View SynthesisAs part of my thesis work, I am working on a new statistical approach for view synthesis. It is particular effective for texture-less regions and specular highlights, two major problems that most existing reconstruction techniques would have difficulty with. We are preparing to report our work to ICCV 2003. Some initial results are presented on the left, the top row shows several input images while the bottom row shows the reconstructed point cloud. |
|
Real-time StereoA multi-resolution stereo algorithm that can be implemented on commodity graphics hardware. A paper will appear in CVPR 2003. We will bring a live demo to CVPR as well. |
|
|
Real-time View Synthesis on Graphics HardwareWe present a novel use of commodity graphics hardware that effectively combines a plane-sweeping algorithm with
view synthesis for real-time, on-line 3D scene acquisition and view synthesis. The heart of our method is to use programmable Pixel
Shader technology to square intensity differences between reference image pixels, and then to choose final colors that correspond to the minimum difference, i.e. the
most consistent color. We filed an invention disclosure with UNC. |
|
Eye-Gaze CorrectionMy internship at Microsoft Research (MSR) during summer 2001 has focused on maintaining eye-contact for desktop video teleconferencing. we took a model-based approach that incorporates a detailed individualized three-dimensional head model with stereoscopic analysis. This approach is very effective; we probably achieved the most realistic results in published literature for eye gaze correction (see the images on the left; the middle is the synthesized view that preserves eye-contact, the other two are the input images.). In the process, we can also get very accurate 3D tracking results of the head pose. The images below show the face model projected on the tracked head. MSR has filed two patent applications for our algorithms and systems. |
|
|
Group TeleconferencingWe want to design a system that facilitate many-to-many teleconferencing. Instead of providing a perceptively correct view for every single user, we strive to provide the best approximating view for the entire group as a whole. We demonstrate two real-time acquisition-through-rendering algorithms: one is based on view dependent texture mapping with automatically acquired approximating geometry, and the other uses an array of cameras to perform Light Field style rendering. |
|
3D Tele-ImmersionThe goal of Tele-Immersion is to enable users at geographically distributed sites to collaborate in real time in a shared, simulated environment as if they were in the same physical room.
While the entire project was a interdisciplinary, multi-site
collaboration, I was mainly invovled in in real-time data capture and
distribution. |
|
2D Immersive TeleconferencingWe worked on improving the field of view and resolution for 2D video teleconferencing. The result is a simple, yet effective technique for producing geometrically correct imagery for teleconferencing environments. The necessary image transformations are derived by finding a direct one-to-one mapping between a capture device and a display device for a fixed viewer location, thus completely avoiding the need for any intermediate, complex representations of screen geometry, capture and display distortions, and viewer location. Using this technique, we can easily build an immersive teleconferencing system using multiple projectors and cameras. |
|
PixelFlex: A Reconfigurable Multi-Projector Display SystemThe PixelFlex system is composed of ceiling-mounted projectors, each with |
Automatic Projector Display Surface Estimation Using Every-Day ImageryWe introduce a new method for continuous display surface auto-calibration. Using a camera that observes the display surface, we match image features in whatever imagery is being projected, with the corresponding features that appear on the display surface, to continually refine an estimate for the display surface geometry. In effect we enjoy the high signal-to-noise ratio of "structured" light (without getting to choose the structure) and the unobtrusive nature of passive correlation-based methods.
|