66 research outputs found
Decoupled Sampling for Graphics Pipelines
We propose a generalized approach to decoupling shading from visibility sampling in graphics pipelines, which we call decoupled sampling. Decoupled sampling enables stochastic supersampling of motion and defocus blur at reduced shading cost, as well as controllable or adaptive shading rates which trade off shading quality for performance. It can be thought of as a generalization of multisample antialiasing (MSAA) to support complex and dynamic mappings from visibility to shading samples, as introduced by motion and defocus blur and adaptive shading. It works by defining a many-to-one hash from visibility to shading samples, and using a buffer to memoize shading samples and exploit reuse across visibility samples. Decoupled sampling is inspired by the Reyes rendering architecture, but like traditional graphics pipelines, it shades fragments rather than micropolygon vertices, decoupling shading from the geometry sampling rate. Also unlike Reyes, decoupled sampling only shades fragments after precise computation of visibility, reducing overshading.
We present extensions of two modern graphics pipelines to support decoupled sampling: a GPU-style sort-last fragment architecture, and a Larrabee-style sort-middle pipeline. We study the architectural implications of decoupled sampling and blur, and derive end-to-end performance estimates on real applications through an instrumented functional simulator. We demonstrate high-quality motion and defocus blur, as well as variable and adaptive shading rates
Decoupled Sampling for Real-Time Graphics Pipelines
We propose decoupled sampling, an approach that decouples shading from visibility sampling in order to enable motion blur and depth-of-field at reduced cost. More generally, it enables extensions of modern real-time graphics pipelines that provide controllable shading rates to trade off quality for performance. It can be thought of as a generalization of GPU-style multisample antialiasing (MSAA) to support unpredictable shading rates, with arbitrary mappings from visibility to shading samples as introduced by motion blur, depth-of-field, and adaptive shading. It is inspired by the Reyes architecture in offline rendering, but targets real-time pipelines by driving shading from visibility samples as in GPUs, and removes the need for micropolygon dicing or rasterization. Decoupled Sampling works by defining a many-to-one hash from visibility to shading samples, and using a buffer to memoize shading samples and exploit reuse across visibility samples. We present extensions of two modern GPU pipelines to support decoupled sampling: a GPU-style sort-last fragment architecture, and a Larrabee-style sort-middle pipeline. We study the architectural implications and derive end-to-end performance estimates on real applications through an instrumented functional simulator. We demonstrate high-quality motion blur and depth-of-field, as well as variable and adaptive shading rates
Discontinuity Edge Overdraw
Aliasing is an important problem when rendering triangle meshes. Efficient antialiasing techniques such as mipmapping greatly improve the filtering of textures defined over a mesh. A major component of the remaining aliasing occurs along discontinuity edges such as silhouettes, creases, and material boundaries. Framebuffer supersampling is a simple remedy, but 2x2 supersampling leaves behind significant temporal artifacts, while greater supersampling demands even more fill-rate and memory. We present an alternative that focuses effort on discontinuity edges by overdrawing such edges as antialiased lines. Although the idea is simple, several subtleties arise. Visible silhouette edges must be detected efficiently. Discontinuity edges need consistent orientations. They must be blended as they approach the silhouette to avoid popping. Unfortunately, edge blending results in blurriness. Our technique balances these two competing objectives of temporal smoothness and spatial sharpness. Finally, the best results are obtained when discontinuity edges are sorted by depth. Our approach proves surprisingly effective at reducing temporal artifacts commonly referred to as "crawling jaggies," with little added cost.Engineering and Applied Science
Image synthesis based on a model of human vision
Modern computer graphics systems are able to construct renderings of such high quality that viewers are deceived into regarding the images as coming from a photographic source. Large amounts of computing resources are expended in this rendering process, using complex mathematical models of lighting and shading.
However, psychophysical experiments have revealed that viewers only regard certain informative regions within a presented image. Furthermore, it has been shown that these visually important regions contain low-level visual feature differences that attract the attention of the viewer.
This thesis will present a new approach to image synthesis that exploits these experimental findings by modulating the spatial quality of image regions by their visual importance. Efficiency gains are therefore reaped, without sacrificing much of the perceived quality of the image. Two tasks must be undertaken to achieve this goal. Firstly, the design of an appropriate region-based model of visual importance, and secondly, the modification of progressive rendering techniques to effect an importance-based rendering approach.
A rule-based fuzzy logic model is presented that computes, using spatial feature differences, the relative visual importance of regions in an image. This model improves upon previous work by incorporating threshold effects induced by global feature difference distributions and by using texture concentration measures.
A modified approach to progressive ray-tracing is also presented. This new approach uses the visual importance model to guide the progressive refinement of an image. In addition, this concept of visual importance has been incorporated into supersampling, texture mapping and computer animation techniques. Experimental results are presented, illustrating the efficiency gains reaped from using this method of progressive rendering.
This visual importance-based rendering approach is expected to have applications in the entertainment industry, where image fidelity may be sacrificed for efficiency purposes, as long as the overall visual impression of the scene is maintained. Different aspects of the approach should find many other applications in image compression, image retrieval, progressive data transmission and active robotic vision
ExWarp: Extrapolation and Warping-based Temporal Supersampling for High-frequency Displays
High-frequency displays are gaining immense popularity because of their
increasing use in video games and virtual reality applications. However, the
issue is that the underlying GPUs cannot continuously generate frames at this
high rate -- this results in a less smooth and responsive experience.
Furthermore, if the frame rate is not synchronized with the refresh rate, the
user may experience screen tearing and stuttering. Previous works propose
increasing the frame rate to provide a smooth experience on modern displays by
predicting new frames based on past or future frames. Interpolation and
extrapolation are two widely used algorithms that predict new frames.
Interpolation requires waiting for the future frame to make a prediction, which
adds additional latency. On the other hand, extrapolation provides a better
quality of experience because it relies solely on past frames -- it does not
incur any additional latency. The simplest method to extrapolate a frame is to
warp the previous frame using motion vectors; however, the warped frame may
contain improperly rendered visual artifacts due to dynamic objects -- this
makes it very challenging to design such a scheme. Past work has used DNNs to
get good accuracy, however, these approaches are slow. This paper proposes
Exwarp -- an approach based on reinforcement learning (RL) to intelligently
choose between the slower DNN-based extrapolation and faster warping-based
methods to increase the frame rate by 4x with an almost negligible reduction in
the perceived image quality
Ray Tracing Gems
This book is a must-have for anyone serious about rendering in real time. With the announcement of new ray tracing APIs and hardware to support them, developers can easily create real-time applications with ray tracing as a core component. As ray tracing on the GPU becomes faster, it will play a more central role in real-time rendering. Ray Tracing Gems provides key building blocks for developers of games, architectural applications, visualizations, and more. Experts in rendering share their knowledge by explaining everything from nitty-gritty techniques that will improve any ray tracer to mastery of the new capabilities of current and future hardware. What you'll learn: The latest ray tracing techniques for developing real-time applications in multiple domains Guidance, advice, and best practices for rendering applications with Microsoft DirectX Raytracing (DXR) How to implement high-performance graphics for interactive visualizations, games, simulations, and more Who this book is for: Developers who are looking to leverage the latest APIs and GPU technology for real-time rendering and ray tracing Students looking to learn about best practices in these areas Enthusiasts who want to understand and experiment with their new GPU
Acceleration Techniques for Photo Realistic Computer Generated Integral Images
The research work presented in this thesis has approached the task of accelerating the
generation of photo-realistic integral images produced by integral ray tracing.
Ray tracing algorithm is a computationally exhaustive algorithm, which spawns one ray
or more through each pixel of the pixels forming the image, into the space containing
the scene. Ray tracing integral images consumes more processing time than normal
images. The unique characteristics of the 3D integral camera model has been analysed
and it has been shown that different coherency aspects than normal ray tracing can be
investigated in order to accelerate the generation of photo-realistic integral images.
The image-space coherence has been analysed describing the relation between rays and
projected shadows in the scene rendered. Shadow cache algorithm has been adapted in
order to minimise shadow intersection tests in integral ray tracing. Shadow intersection
tests make the majority of the intersection tests in ray tracing. Novel pixel-tracing
styles are developed uniquely for integral ray tracing to improve the image-space
coherence and the performance of the shadow cache algorithm. Acceleration of the
photo-realistic integral images generation using the image-space coherence information
between shadows and rays in integral ray tracing has been achieved with up to 41 % of
time saving. Also, it has been proven that applying the new styles of pixel-tracing does
not affect of the scalability of integral ray tracing running over parallel computers.
The novel integral reprojection algorithm has been developed uniquely through
geometrical analysis of the generation of integral image in order to use the tempo-spatial
coherence information within the integral frames. A new derivation of integral
projection matrix for projecting points through an axial model of a lenticular lens has
been established. Rapid generation of 3D photo-realistic integral frames has been
achieved with a speed four times faster than the normal generation
The lightspeed automatic interactive lighting preview system
Thesis (S.M.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2007.Includes bibliographical references (p. 57-59).We present an automated approach for high-quality preview of feature-film rendering during lighting design. Similar to previous work, we use a deep-framebuffer shaded on the GPU to achieve interactive performance. Our first contribution is to generate the deep-framebuffer and corresponding shaders automatically through data-flow analysis and compilation of the original scene. Cache compression reduces automatically-generated deep-framebuffers to reasonable size for complex production scenes and shaders. We also propose a new structure, the indirect framebuffer, that decouples shading samples from final pixels and allows a deep-framebuffer to handle antialiasing, motion blur and transparency efficiently. Progressive refinement enables fast feedback at coarser resolution. We demonstrate our approach in real-world production.by Jonathan Millard Ragan-Kelley.S.M
- …