7 research outputs found
Robust temporal depth enhancement method for dynamic virtual view synthesis
Depth-image-based rendering (DIBR) is a view synthesis technique that generates virtual views by warping from
the reference images based on depth maps. The quality of synthesized views highly depends on the accuracy of
depth maps. However, for dynamic scenarios, depth sequences obtained through stereo matching methods frame
by frame can be temporally inconsistent, especially in static regions, which leads to uncomfortable flickering
artifacts in synthesized videos. This problem can be eliminated by depth enhancement methods that perform
temporal filtering to suppress depth inconsistency, yet those methods may also spread depth errors. Although these
depth enhancement algorithms increase the temporal consistency of synthesized videos, they have the risk of
reducing the quality of rendered videos. Since conventional methods may not achieve both properties, in this paper,
we present for static regions a robust temporal depth enhancement (RTDE) method, which propagates exactly the
reliable depth values into succeeding frames to upgrade not only the accuracy but also the temporal consistency
of depth estimations. This technique benefits the quality of synthesized videos. In addition we propose a novel
evaluation metric to quantitatively compare temporal consistency between our method and the state of arts.
Experimental results demonstrate the robustness of our method for dynamic virtual view synthesis, not only the
temporal consistency but also the quality of synthesized videos in static regions are improved
View Synthesis Tool for VR Immersive Video
This chapter addresses the view synthesis of natural scenes in virtual reality (VR) using depth image-based rendering (DIBR). This method reaches photorealistic results as it directly warps photos to obtain the output, avoiding the need to photograph every possible viewpoint or to make a 3D reconstruction of a scene followed by a ray-tracing rendering. An overview of the DIBR approach and frequently encountered challenges (disocclusion and ghosting artifacts, multi-view blending, handling of non-Lambertian objects) are described. Such technology finds applications in VR immersive displays and holography. Finally, a comprehensive manual of the Reference View Synthesis software (RVS), an open-source tool tested on open datasets and recognized by the MPEG-I standardization activities (where”I″ refers to”immersive”) is described for hands-on practicing
Fusing spatial and temporal components for real-time depth data enhancement of dynamic scenes
The depth images from consumer depth cameras (e.g., structured-light/ToF devices) exhibit a substantial amount of artifacts (e.g., holes, flickering, ghosting) that needs to be removed for real-world applications. Existing methods cannot entirely remove them and perform slow. This thesis proposes a new real-time spatio-temporal depth image enhancement filter that completely removes flickering and ghosting, and significantly reduces holes. This thesis also presents a novel depth-data capture setup and two data reduction methods to optimize the performance of the proposed enhancement method
A descriptive model for determining optimal human performance in systems. Volume 3 - An approach for determining the optimal role of man and allocation of functions in an aerospace system
Optimal role of man in space, allocation of men and machines in aerospace systems, and descriptive model for determining optimal human performanc