7,068 research outputs found

    Soft bilateral filtering shadows using multiple image-based algorithms

    Get PDF
    This study introduces Soft Bilateral Filtering Shadows method of dynamic scenes, which uses multi-matrices of the light sample points due to lack realism in soft shadows generation in real time. While geometry-based shadow algorithm requires one pass per polygon for rendering shadow that requires time-consuming, the adopted shadow map algorithm needs a single rendering pass for each sample point of the light source to generate shadow at low cost. This method renders a complex scenes and accurately eliminating the inherent deficiencies in shadow maps. In order to compute shadow maps, view matrices were used for each sample point of the extended light source. Then penumbra region was used for interpolation based on bilateral filtering to create the soft shadows. They depend on multiple shadow maps which provide antialiasing shadow maps. The method uses fragment shader for rendering multiple shadow maps with penumbra and umbra regions. The main contribution of this article is introducing interpolation bilaterally of image-based shadows. This method makes the most effect of the computation significantly appear at the edges of the penumbra region. Furthermore, the filtering allows to obtain on the soft shadow marvelously at the lowest number possible of the light sample points. The generated soft shadows have good performance and high quality therefore, they are suitable for interactive applications. © 2016 Springer Science+Business Media New Yor

    ITEM: Inter-Texture Error Measurement for 3D Meshes

    Get PDF
    We introduce a simple and innovative method to compare any two texture maps, regardless of their sizes, aspect ratios, or even masks, as long as they are both meant to be mapped onto the same 3D mesh. Our system is based on a zero-distortion 3D mesh unwrapping technique which compares two new adapted texture atlases with the same mask but different texel colors, and whose every texel covers the same area in 3D. Once these adapted atlases are created, we measure their difference with ITEM-RMSE, a slightly modified version of the standard RMSE defined for images. ITEM-RMSE is more meaningful and reliable than RMSE because it only takes into account the texels inside the mask, since they are the only ones that will actually be used during rendering. Our method is not only very useful to compare the space efficiency of different texture atlas generation algorithms, but also to quantify texture loss in compression schemes for multi-resolution textured 3D meshes

    MVF-Net: Multi-View 3D Face Morphable Model Regression

    Full text link
    We address the problem of recovering the 3D geometry of a human face from a set of facial images in multiple views. While recent studies have shown impressive progress in 3D Morphable Model (3DMM) based facial reconstruction, the settings are mostly restricted to a single view. There is an inherent drawback in the single-view setting: the lack of reliable 3D constraints can cause unresolvable ambiguities. We in this paper explore 3DMM-based shape recovery in a different setting, where a set of multi-view facial images are given as input. A novel approach is proposed to regress 3DMM parameters from multi-view inputs with an end-to-end trainable Convolutional Neural Network (CNN). Multiview geometric constraints are incorporated into the network by establishing dense correspondences between different views leveraging a novel self-supervised view alignment loss. The main ingredient of the view alignment loss is a differentiable dense optical flow estimator that can backpropagate the alignment errors between an input view and a synthetic rendering from another input view, which is projected to the target view through the 3D shape to be inferred. Through minimizing the view alignment loss, better 3D shapes can be recovered such that the synthetic projections from one view to another can better align with the observed image. Extensive experiments demonstrate the superiority of the proposed method over other 3DMM methods.Comment: 2019 Conference on Computer Vision and Pattern Recognitio

    Object-based 2D-to-3D video conversion for effective stereoscopic content generation in 3D-TV applications

    Get PDF
    Three-dimensional television (3D-TV) has gained increasing popularity in the broadcasting domain, as it enables enhanced viewing experiences in comparison to conventional two-dimensional (2D) TV. However, its application has been constrained due to the lack of essential contents, i.e., stereoscopic videos. To alleviate such content shortage, an economical and practical solution is to reuse the huge media resources that are available in monoscopic 2D and convert them to stereoscopic 3D. Although stereoscopic video can be generated from monoscopic sequences using depth measurements extracted from cues like focus blur, motion and size, the quality of the resulting video may be poor as such measurements are usually arbitrarily defined and appear inconsistent with the real scenes. To help solve this problem, a novel method for object-based stereoscopic video generation is proposed which features i) optical-flow based occlusion reasoning in determining depth ordinal, ii) object segmentation using improved region-growing from masks of determined depth layers, and iii) a hybrid depth estimation scheme using content-based matching (inside a small library of true stereo image pairs) and depth-ordinal based regularization. Comprehensive experiments have validated the effectiveness of our proposed 2D-to-3D conversion method in generating stereoscopic videos of consistent depth measurements for 3D-TV applications
    corecore