86 research outputs found

    Efficient Many-Light Rendering of Scenes with Participating Media

    Get PDF
    We present several approaches based on virtual lights that aim at capturing the light transport without compromising quality, and while preserving the elegance and efficiency of many-light rendering. By reformulating the integration scheme, we obtain two numerically efficient techniques; one tailored specifically for interactive, high-quality lighting on surfaces, and one for handling scenes with participating media

    Acceleration Techniques for Photo Realistic Computer Generated Integral Images

    Get PDF
    The research work presented in this thesis has approached the task of accelerating the generation of photo-realistic integral images produced by integral ray tracing. Ray tracing algorithm is a computationally exhaustive algorithm, which spawns one ray or more through each pixel of the pixels forming the image, into the space containing the scene. Ray tracing integral images consumes more processing time than normal images. The unique characteristics of the 3D integral camera model has been analysed and it has been shown that different coherency aspects than normal ray tracing can be investigated in order to accelerate the generation of photo-realistic integral images. The image-space coherence has been analysed describing the relation between rays and projected shadows in the scene rendered. Shadow cache algorithm has been adapted in order to minimise shadow intersection tests in integral ray tracing. Shadow intersection tests make the majority of the intersection tests in ray tracing. Novel pixel-tracing styles are developed uniquely for integral ray tracing to improve the image-space coherence and the performance of the shadow cache algorithm. Acceleration of the photo-realistic integral images generation using the image-space coherence information between shadows and rays in integral ray tracing has been achieved with up to 41 % of time saving. Also, it has been proven that applying the new styles of pixel-tracing does not affect of the scalability of integral ray tracing running over parallel computers. The novel integral reprojection algorithm has been developed uniquely through geometrical analysis of the generation of integral image in order to use the tempo-spatial coherence information within the integral frames. A new derivation of integral projection matrix for projecting points through an axial model of a lenticular lens has been established. Rapid generation of 3D photo-realistic integral frames has been achieved with a speed four times faster than the normal generation

    Hierarchical N-Body problem on graphics processor unit

    Get PDF
    Galactic simulation is an important cosmological computation, and represents a classical N-body problem suitable for implementation on vector processors. Barnes-Hut algorithm is a hierarchical N-Body method used to simulate such galactic evolution systems. Stream processing architectures expose data locality and concurrency available in multimedia applications. On the other hand, there are numerous compute-intensive scientific or engineering applications that can potentially benefit from such computational and communication models. These applications are traditionally implemented on vector processors. Stream architecture based graphics processor units (GPUs) present a novel computational alternative for efficiently implementing such high-performance applications. Rendering on a stream architecture sustains high performance, while user-programmable modules allow implementing complex algorithms efficiently. GPUs have evolved over the years, from being fixed-function pipelines to user programmable processors. In this thesis, we focus on the implementation of Barnes-Hut algorithm on typical current-generation programmable GPUs. We exploit computation and communication requirements present in Barnes-Hut algorithm to expose their suitability for user-programmable GPUs. Our implementation of the Barnes-Hut algorithm is formulated as a fragment shader targeting the selected GPU. We discuss implementation details, design issues, results, and challenges encountered in programming the fragment shader

    Shadow Computations using Robust Epsilon Visibility

    Get PDF
    Analytic visibility algorithms, for example methods which compute a subdivided mesh to represent shadows, are notoriously unrobust and hard to use in practice. We present a new method based on a generalized definition of extremal stabbing lines, which are the extremities of shadow boundaries. We treat scenes containing multiple edges or vertices in degenerate configurations, (e.g., collinear or coplanar). We introduce a robust epsilon method to determine whether each generalized extremal stabbing line is blocked, or is touched by these scene elements, and thus added to the line's generators. We develop robust blocker predicates for polygons which are smaller than epsilon. For larger values, small shadow features merge and eventually disappear. We can thus robustly connect generalized extremal stabbing lines in degenerate scenes to form shadow boundaries. We show that our approach is consistent, and that shadow boundary connectivity is preserved when features merge. We have implemented our algorithm, and show that we can robustly compute analytic shadow boundaries to the precision of our chosen epsilon threshold for non-trivial models, containing numerous degeneracies

    The Visibility Skeleton: A Powerful and Multi-Purpose Global Visibility Tool

    Get PDF
    International audienceMany problems in computer graphics and computer vision require accurate global visibility information. Previous approaches have typically been complicated to implement and numerically unstable, and often too expensive in storage or computation. The Visibility Skeleton is a new powerful utility which can efficiently and accurately answer visibility queries for the entire scene. The Visibility Skeleton is a multi-purpose tool, which can solve numerous different problems. A simple construction algorithm is presented which only requires the use of well known computer graphics algorithmic components such as ray-casting and line/plane intersections. We provide an exhaustive catalogue of visual events which completely encode all possible visibility changes of a polygonal scene into a graph structure. The nodes of the graph are extremal stabbing lines, and the arcs are critical line swaths. Our implementation demonstrates the construction of the Visibility Skeleton for scenes of over a thousand polygons. We also show its use to compute exact visible boundaries of a vertex with respect to any polygon in the scene, the computation of global or on-the-fly discontinuity meshes by considering any scene polygon as a source, as well as the extraction of the exact blocker list between any polygon pair. The algorithm is shown to be manageable for the scenes tested, both in storage and in computation time. To address the potential complexity problems for large scenes, on-demand or lazy contruction is presented, its implementation showing encouraging first results

    Second-Order Occlusion-Aware Volumetric Radiance Caching

    Get PDF
    We present a second-order gradient analysis of light transport in participating media and use this to develop an improved radiance caching algorithm for volumetric light transport. We adaptively sample and interpolate radiance from sparse points in the medium using a second-order Hessian-based error metric to determine when interpolation is appropriate. We derive our metric from each point's incoming light field, computed by using a proxy triangulation-based representation of the radiance reflected by the surrounding medium and geometry. We use this representation to efficiently compute the first- and second-order derivatives of the radiance at the cache points while accounting for occlusion changes. We also propose a self-contained two-dimensional model for light transport in media and use it to validate and analyze our approach, demonstrating that our method outperforms previous radiance caching algorithms both in terms of accurate derivative estimates and final radiance extrapolation. We generalize these findings to practical three-dimensional scenarios, where we show improved results while reducing computation time by up to 30\% compared to previous work

    Path manipulation strategies for rendering dynamic environments.

    Get PDF
    The current work introduces path manipulation as a tool that extends bidirectional path tracing to reuse paths in the temporal domain. Defined as an apparatus of sampling and reuse strategies, path manipulation reconstructs the subpaths that compose the light transport paths and addresses the restriction of static geometry commonly associated with Monte Carlo light transport simulations. By reconstructing and reusing subpaths, the path manipulation algorithm obviates the regeneration of the entire path collection, reduces the computational load of the original algorithm and supports scene dynamism. Bidirectional path tracing relies on local path sampling techniques to generate the paths of light in a synthetic environment. By using the information localized at path vertices, like the probability distribution, the sampling techniques construct paths progressively with distinct probability densities. Each probability density corresponds to a particular sampling technique, which accounts for specific illumination effects. Bidirectional path tracing uses multiple importance sampling to combine paths sampled with different techniques in low-variance estimators. The path sampling techniques and multiple importance sampling are the keys to the efficacy of bidirectional path tracing. However, the sampling techniques gained little attention beyond the generation and evaluation of paths. Bidirectional path tracing was designed for static scenes and thus it discards the generated paths immediately after the evaluation of their contributions. Limiting the lifespan of paths to a generation-evaluation cycle imposes a static use of paths and of sampling techniques. The path manipulation algorithm harnesses the potential of the sampling techniques to supplant the static manipulation of paths with a generation-evaluation-reuse cycle. An intra-subpath connectivity strategy was devised to reconnect the segregated chains of the subpaths invalidated by the scene alterations. Successful intra-subpath connections generate subpaths in multiple pieces by reusing subpath chains from prior frames. Subpaths are reconstructed generically, regardless of the subpath or scene dynamism type and without the need for predefined animation paths. The result is the extension of bidirectional path tracing to the temporal domain

    Towards Predictive Rendering in Virtual Reality

    Get PDF
    The strive for generating predictive images, i.e., images representing radiometrically correct renditions of reality, has been a longstanding problem in computer graphics. The exactness of such images is extremely important for Virtual Reality applications like Virtual Prototyping, where users need to make decisions impacting large investments based on the simulated images. Unfortunately, generation of predictive imagery is still an unsolved problem due to manifold reasons, especially if real-time restrictions apply. First, existing scenes used for rendering are not modeled accurately enough to create predictive images. Second, even with huge computational efforts existing rendering algorithms are not able to produce radiometrically correct images. Third, current display devices need to convert rendered images into some low-dimensional color space, which prohibits display of radiometrically correct images. Overcoming these limitations is the focus of current state-of-the-art research. This thesis also contributes to this task. First, it briefly introduces the necessary background and identifies the steps required for real-time predictive image generation. Then, existing techniques targeting these steps are presented and their limitations are pointed out. To solve some of the remaining problems, novel techniques are proposed. They cover various steps in the predictive image generation process, ranging from accurate scene modeling over efficient data representation to high-quality, real-time rendering. A special focus of this thesis lays on real-time generation of predictive images using bidirectional texture functions (BTFs), i.e., very accurate representations for spatially varying surface materials. The techniques proposed by this thesis enable efficient handling of BTFs by compressing the huge amount of data contained in this material representation, applying them to geometric surfaces using texture and BTF synthesis techniques, and rendering BTF covered objects in real-time. Further approaches proposed in this thesis target inclusion of real-time global illumination effects or more efficient rendering using novel level-of-detail representations for geometric objects. Finally, this thesis assesses the rendering quality achievable with BTF materials, indicating a significant increase in realism but also confirming the remainder of problems to be solved to achieve truly predictive image generation

    Efficient global illumination calculation for inverse lighting problems

    Get PDF
    La luz es un elemento clave en la manera en que percibimos y experimentamos nuestro entorno. Como tal, es un objeto mas a modelar en el proceso de diseño, de forma similar a como ocurre con las formas y los materiales. Las intenciones de iluminacion (LI) son los objetivos y restricciones que el diseñador pretende alcanzar en el proceso del diseño de iluminaci´on: ¿qué superficies se deben iluminar con luz natural y cuales con luz artificial?, ¿qué zonas deben estar en sombra?, ¿cuales son las intensidades maximas y mínimas permitidas? Satisfacer las LI consiste en encontrar la ubicacion, forma e intensidad adecuada de las fuentes luminosas. Este tipo de problemas se define como un problema inverso de iluminacion (ILP) que se resuelve con tecnicas de optimizacion. En el contexto anterior, el objetivo de esta tesis consiste en proponer metodos eficientes para resolver ILP. Este objetivo es motivado por la brecha percibida entre los problemas habituales de diseño de iluminacion y las herramientas computacionales existentes para su resolucion. Las herramientas desarrolladas por la industria se especializan en evaluar configuraciones de iluminacion previamente diseñadas, y las desarrolladas por la academia resuelven problemas relativamente sencillos a costos elevados. Las propuestas cubren distintos aspectos del proceso de optimizacion, que van desde la formulacion del problema a su resolucion. Estan desarrolladas para el caso en que las superficies poseen reflexion e iluminacion difusas y se basan en el calculo de una aproximacion de rango bajo de la matriz de radiosidad. Algunos resultados obtenidos son: el calculo acelerado de la radiosidad de la escena en una unidad de procesamiento gr´afico (GPU); el uso de la heuristica \201Cvariable neighborhood search\201D (VNS) para la resolucion de ILP; el planteo de una estructura multinivel para tratar ILP de forma escalonada; y el uso de tecnicas para optimizar la configuracion de filtros de luz. Otros resultados obtenidos se basan en la formulacion de las LI en funcion de la media y desviacion estandar de las radiosidades halladas. Se propone un metodo para generar LI que contengan esos parametros estadisticos, y otro metodo para acelerar su evaluacion. Con estos resultados se logran tiempos de respuesta interactivos. Por último, las tecnicas anteriores adolecen de una etapa de pre-cómputo relativamente costosa, por tanto se propone acelerar el calculo de la inversa de la matriz de radiosidad a partir de una muestra de factores de forma. Los métodos aquí presentados fueron publicados en seis articulos, tres de ellos en congresos internacionales y tres en revistas arbitradas.Light is a key element that influences the way we perceive and experience our environment. As such, light is an object to be modeled in the design process, as happens with the forms and materials. The lighting intentions (LI) are the objectives and constraints that designers want to achieve in the process of lighting design: which surfaces should be illuminated with natural and which with artificial light?, which surfaces should be in shadow?, which are the maximum and minimum intensities allowed? The fulfillment of the LI consists in finding the location, shape and intensity appropriate for the light sources. This problem is defined as an inverse lighting problem (ILP), solved by optimization techniques. In the above context, the aim of this thesis is the proposal of efficient methods to solve ILP. This objective is motivated by the perceived gap between the usual problems of lighting design, and the computational tools developed for its resolution. The tools developed by the industry specialize in evaluating previously designed lighting configurations, and those developed by the academia solve relatively simple problems at a high computational cost. The proposals cover several aspects of the optimization process, ranging from the formulation of the problem to its resolution. They are developed for the case in which the surfaces have Lambertian reflection and illumination, and are based on the calculation of a low rank approximation to the radiosity matrix. Some results are: rapid calculation of radiosity of the scene in a graphics processing unit (GPU), the use of heuristics “variable neighborhood search” (VNS) for solving ILP, the proposition of a multilevel structure to solve ILP in a stepwise approach, and the use of these techniques to optimize the configuration of light filters. Other results are based on the formulation of LI that use the mean and standard deviation of the radiosity values found. A method is proposed for generating LI containing these parameters, and another method is developed to speed up their evaluations. With these results we achieve interactive response times. Finally, the above techniques suffer from a costly pre-computing stage and therefore, a method is proposed to accelerate the calculation of the radiosity inverse matrix based on a sample of the form factors. The methods presented here were published in six articles, three of them at international conferences and three in peer reviewed journals
    • …
    corecore