153,957 research outputs found
Stochastic Modeling and Resolution-Free Rendering of Film Grain
The realistic synthesis and rendering of film grain is a crucial goal for many amateur and professional photographers and film-makers whose artistic works require the authentic feel of analog photography. The objective of this work is to propose an algorithm that reproduces the visual aspect of film grain texture on any digital image. Previous approaches to this problem either propose unrealistic models or simply blend scanned images of film grain with the digital image, in which case the result is inevitably limited by the quality and resolution of the initial scan. In this work, we introduce a stochastic model to approximate the physical reality of film grain, and propose a resolution-free rendering algorithm to simulate realistic film grain for any digital input image. By varying the parameters of this model, we can achieve a wide range of grain types. We demonstrate this by comparing our results with film grain examples from dedicated software, and show that our rendering results closely resemble these real film emulsions. In addition to realistic grain rendering, our resolution-free algorithm allows for any desired zoom factor, even down to the scale of the microscopic grains themselves
Virtual Rephotography: Novel View Prediction Error for 3D Reconstruction
The ultimate goal of many image-based modeling systems is to render
photo-realistic novel views of a scene without visible artifacts. Existing
evaluation metrics and benchmarks focus mainly on the geometric accuracy of the
reconstructed model, which is, however, a poor predictor of visual accuracy.
Furthermore, using only geometric accuracy by itself does not allow evaluating
systems that either lack a geometric scene representation or utilize coarse
proxy geometry. Examples include light field or image-based rendering systems.
We propose a unified evaluation approach based on novel view prediction error
that is able to analyze the visual quality of any method that can render novel
views from input images. One of the key advantages of this approach is that it
does not require ground truth geometry. This dramatically simplifies the
creation of test datasets and benchmarks. It also allows us to evaluate the
quality of an unknown scene during the acquisition and reconstruction process,
which is useful for acquisition planning. We evaluate our approach on a range
of methods including standard geometry-plus-texture pipelines as well as
image-based rendering techniques, compare it to existing geometry-based
benchmarks, and demonstrate its utility for a range of use cases.Comment: 10 pages, 12 figures, paper was submitted to ACM Transactions on
Graphics for revie
Image-based rendering and synthesis
Multiview imaging (MVI) is currently the focus of some research as it has a wide range of applications and opens up research in other topics and applications, including virtual view synthesis for three-dimensional (3D) television (3DTV) and entertainment. However, a large amount of storage is needed by multiview systems and are difficult to construct. The concept behind allowing 3D scenes and objects to be visualized in a realistic way without full 3D model reconstruction is image-based rendering (IBR). Using images as the primary substrate, IBR has many potential applications including for video games, virtual travel and others. The technique creates new views of scenes which are reconstructed from a collection of densely sampled images or videos. The IBR concept has different classification such as knowing 3D models and the lighting conditions and be rendered using conventional graphic techniques. Another is lightfield or lumigraph rendering which depends on dense sampling with no or very little geometry for rendering without recovering the exact 3D-models.published_or_final_versio
P-Buffer: Hidden-line rendering with a dynamic p-buffer
Despite the emergence of highly realistic computer-generated images, line-drawing images are still a common practice in showing the shapes and movements of three-dimensional objects. It is especially true when rendering time is critical in interactive applications such as the modeling and testing stage of computer aided design/manufacturing, computer animation, and virtual reality. Hence much effort has been devoted to provide sufficient information of the displayed objects with the least amount of time. While the techniques that determine visible surfaces in an image-space have the advantages on rendering speed and processable shapes, those that decide visible lines or line segments in an object-space are more suitable for showing hidden lines. The P-buffer algorithm introduced in this paper is a method for rendering line-drawing images with dashed hidden-lines. Being an image-space method, this algorithm preserves the low computational cost and works on a wide range of object shapes; as an extension to the Z-buffer algorithm it, moreover, discloses hidden surfaces by showing them with dashed lines. After a discussion on rendering techniques of line-drawing images, this paper presents this algorithm with pseudocode in C++ programming language and shows some experimental results as well. This image-space algorithm can be used as a compromise approach that reveals the concealed information of hidden-surface-removed views for time-critical rendering
- …