5,604 research outputs found

    A Survey of Ocean Simulation and Rendering Techniques in Computer Graphics

    Get PDF
    This paper presents a survey of ocean simulation and rendering methods in computer graphics. To model and animate the ocean's surface, these methods mainly rely on two main approaches: on the one hand, those which approximate ocean dynamics with parametric, spectral or hybrid models and use empirical laws from oceanographic research. We will see that this type of methods essentially allows the simulation of ocean scenes in the deep water domain, without breaking waves. On the other hand, physically-based methods use Navier-Stokes Equations (NSE) to represent breaking waves and more generally ocean surface near the shore. We also describe ocean rendering methods in computer graphics, with a special interest in the simulation of phenomena such as foam and spray, and light's interaction with the ocean surface

    Fluids real-time rendering

    Get PDF
    In this thesis the existing methods for realistic visualization of uids in real-time are reviewed. The correct handling of the interaction of light with a uid surface can highly increase the realism of the rendering, therefore method for physically accurate rendering of re ections and refractions will be used. The light- uid interaction does not stop at the surface, but continues inside the uid volume, causing caustics and beams of light. The simulation of uids require extremely time-consuming processes to achieve physical accuracy and will not be explored, although the main concepts will be given. Therefore, the main goals of this work are: Study and review the existing methods for rendering uids in realtime. Find a simpli ed physical model of light interaction, because a complete physically correct model would not achieve real-time. Develop an application that uses the found methods and the light interaction model

    NaRPA: Navigation and Rendering Pipeline for Astronautics

    Full text link
    This paper presents Navigation and Rendering Pipeline for Astronautics (NaRPA) - a novel ray-tracing-based computer graphics engine to model and simulate light transport for space-borne imaging. NaRPA incorporates lighting models with attention to atmospheric and shading effects for the synthesis of space-to-space and ground-to-space virtual observations. In addition to image rendering, the engine also possesses point cloud, depth, and contour map generation capabilities to simulate passive and active vision-based sensors and to facilitate the designing, testing, or verification of visual navigation algorithms. Physically based rendering capabilities of NaRPA and the efficacy of the proposed rendering algorithm are demonstrated using applications in representative space-based environments. A key demonstration includes NaRPA as a tool for generating stereo imagery and application in 3D coordinate estimation using triangulation. Another prominent application of NaRPA includes a novel differentiable rendering approach for image-based attitude estimation is proposed to highlight the efficacy of the NaRPA engine for simulating vision-based navigation and guidance operations.Comment: 49 pages, 22 figure

    Efficient Methods for Computational Light Transport

    Get PDF
    En esta tesis presentamos contribuciones sobre distintos retos computacionales relacionados con transporte de luz. Los algoritmos que utilizan información sobre el transporte de luz están presentes en muchas aplicaciones de hoy en día, desde la generación de efectos visuales, a la detección de objetos en tiempo real. La luz es una valiosa fuente de información que nos permite entender y representar nuestro entorno, pero obtener y procesar esta información presenta muchos desafíos debido a la complejidad de las interacciones entre la luz y la materia. Esta tesis aporta contribuciones en este tema desde dos puntos de vista diferentes: algoritmos en estado estacionario, en los que se asume que la velocidad de la luz es infinita; y algoritmos en estado transitorio, que tratan la luz no solo en el dominio espacial, sino también en el temporal. Nuestras contribuciones en algoritmos estacionarios abordan problemas tanto en renderizado offline como en tiempo real. Nos enfocamos en la reducción de varianza para métodos offline,proponiendo un nuevo método para renderizado eficiente de medios participativos. En renderizado en tiempo real, abordamos las limitacionesde consumo de batería en dispositivos móviles proponiendo un sistema de renderizado que incrementa la eficiencia energética en aplicaciones gráficas en tiempo real. En el transporte de luz transitorio, formalizamos la simulación de este tipo transporte en este nuevo dominio, y presentamos nuevos algoritmos y métodos para muestreo eficiente para render transitorio. Finalmente, demostramos la utilidad de generar datos en este dominio, presentando un nuevo método para corregir interferencia multi-caminos en camaras Timeof- Flight, un problema patológico en el procesamiento de imágenes transitorias.n this thesis we present contributions to different challenges of computational light transport. Light transport algorithms are present in many modern applications, from image generation for visual effects to real-time object detection. Light is a rich source of information that allows us to understand and represent our surroundings, but obtaining and processing this information presents many challenges due to its complex interactions with matter. This thesis provides advances in this subject from two different perspectives: steady-state algorithms, where the speed of light is assumed infinite, and transient-state algorithms, which deal with light as it travels not only through space but also time. Our steady-state contributions address problems in both offline and real-time rendering. We target variance reduction in offline rendering by proposing a new efficient method for participating media rendering. In real-time rendering, we target energy constraints of mobile devices by proposing a power-efficient rendering framework for real-time graphics applications. In transient-state we first formalize light transport simulation under this domain, and present new efficient sampling methods and algorithms for transient rendering. We finally demonstrate the potential of simulated data to correct multipath interference in Time-of-Flight cameras, one of the pathological problems in transient imaging.<br /

    Quantitative Estimation of Surface Soil Moisture in Agricultural Landscapes using Spaceborne Synthetic Aperture Radar Imaging at Different Frequencies and Polarizations

    Get PDF
    Soil moisture and its distribution in space and time plays an important role in the surface energy balance at the soil-atmosphere interface. It is a key variable influencing the partitioning of solar energy into latent and sensible heat flux as well as the partitioning of precipitation into runoff and percolation. Due to their large spatial variability, estimation of spatial patterns of soil moisture from field measurements is difficult and not feasible for large scale analyses. In the past decades, Synthetic Aperture Radar (SAR) remote sensing has proven its potential to quantitatively estimate near surface soil moisture at high spatial resolutions. Since the knowledge of the basic SAR concepts is important to understand the impact of different natural terrain features on the quantitative estimation of soil moisture and other surface parameters, the fundamental principles of synthetic aperture radar imaging are discussed. Also the two spaceborne SAR missions whose data was used in this study, the ENVISAT of the European Space Agency (ESA) and the ALOS of the Japanese Aerospace Exploration Agency (JAXA), are introduced. Subsequently, the two essential surface properties in the field of radar remote sensing, surface soil moisture and surface roughness are defined, and the established methods of their measurement are described. The in situ data used in this study, as well as the research area, the River Rur catchment, with the individual test sites where the data was collected between 2007 and 2010, are specified. On this basis, the important scattering theories in radar polarimetry are discussed and their application is demonstrated using novel polarimetric ALOS/PALSAR data. A critical review of different classical approaches to invert soil moisture from SAR imaging is provided. Five prevalent models have been chosen with the aim to provide an overview of the evolution of ideas and techniques in the field of soil moisture estimation from active microwave data. As the core of this work, a new semi-empirical model for the inversion of surface soil moisture from dual polarimetric L-band SAR data is introduced. This novel approach utilizes advanced polarimetric decomposition techniques to correct for the disturbing effects from surface roughness and vegetation on the soil moisture retrieval without the use of a priori knowledge. The land use specific algorithms for bare soil, grassland, sugar beet, and winter wheat allow quantitative estimations with accuracies in the order of 4 Vol.-%. Application of remotely sensed soil moisture patterns is demonstrated on the basis of mesoscale SAR data by investigating the variability of soil moisture patterns at different spatial scales ranging from field scale to catchment scale. The results show that the variability of surface soil moisture decreases with increasing wetness states at all scales. Finally, the conclusions from this dissertational research are summarized and future perspectives on how to extend the proposed model by means of improved ground based measurements and upcoming advances in sensor technology are discussed. The results obtained in this thesis lead to the conclusion that state-of-the-art spaceborne dual polarimetric L-band SAR systems are not only suitable to accurately retrieve surface soil moisture contents of bare as well as of vegetated agricultural fields and grassland, but for the first time also allow investigating within-field spatial heterogeneities from space

    Enhanced perception in volume visualization

    Get PDF
    Due to the nature of scientic data sets, the generation of convenient visualizations may be a difficult task, but crucial to correctly convey the relevant information of the data. When working with complex volume models, such as the anatomical ones, it is important to provide accurate representations, since a misinterpretation can lead to serious mistakes while diagnosing a disease or planning surgery. In these cases, enhancing the perception of the features of interest usually helps to properly understand the data. Throughout years, researchers have focused on different methods to improve the visualization of volume data sets. For instance, the definition of good transfer functions is a key issue in Volume Visualization, since transfer functions determine how materials are classified. Other approaches are based on simulating realistic illumination models to enhance the spatial perception, or using illustrative effects to provide the level of abstraction needed to correctly interpret the data. This thesis contributes with new approaches to enhance the visual and spatial perception in Volume Visualization. Thanks to the new computing capabilities of modern graphics hardware, the proposed algorithms are capable of modifying the illumination model and simulating illustrative motifs in real time. In order to enhance local details, which are useful to better perceive the shape and the surfaces of the volume, our first contribution is an algorithm that employs a common sharpening operator to modify the lighting applied. As a result, the overall contrast of the visualization is enhanced by brightening the salient features and darkening the deeper regions of the volume model. The enhancement of depth perception in Direct Volume Rendering is also covered in the thesis. To do this, we propose two algorithms to simulate ambient occlusion: a screen-space technique based on using depth information to estimate the amount of light occluded, and a view-independent method that uses the density values of the data set to estimate the occlusion. Additionally, depth perception is also enhanced by adding halos around the structures of interest. Maximum Intensity Projection images provide a good understanding of the high intensity features of the data, but lack any contextual information. In order to enhance the depth perception in such a case, we present a novel technique based on changing how intensity is accumulated. Furthermore, the perception of the spatial arrangement of the displayed structures is also enhanced by adding certain colour cues. The last contribution is a new manipulation tool designed for adding contextual information when cutting the volume. Based on traditional illustrative effects, this method allows the user to directly extrude structures from the cross-section of the cut. As a result, the clipped structures are displayed at different heights, preserving the information needed to correctly perceive them.Debido a la naturaleza de los datos científicos, visualizarlos correctamente puede ser una tarea complicada, pero crucial para interpretarlos de forma adecuada. Cuando se trabaja con modelos de volumen complejos, como es el caso de los modelos anatómicos, es importante generar imágenes precisas, ya que una mala interpretación de las mismas puede producir errores graves en el diagnóstico de enfermedades o en la planificación de operaciones quirúrgicas. En estos casos, mejorar la percepción de las zonas de interés, facilita la comprensión de la información inherente a los datos. Durante décadas, los investigadores se han centrado en el desarrollo de técnicas para mejorar la visualización de datos volumétricos. Por ejemplo, los métodos que permiten definir buenas funciones de transferencia son clave, ya que éstas determinan cómo se clasifican los materiales. Otros ejemplos son las técnicas que simulan modelos de iluminación realista, que permiten percibir mejor la distribución espacial de los elementos del volumen, o bien los que imitan efectos ilustrativos, que proporcionan el nivel de abstracción necesario para interpretar correctamente los datos. El trabajo presentado en esta tesis se centra en mejorar la percepción de los elementos del volumen, ya sea modificando el modelo de iluminación aplicado en la visualización, o simulando efectos ilustrativos. Aprovechando la capacidad de cálculo de los nuevos procesadores gráficos, se describen un conjunto de algoritmos que permiten obtener los resultados en tiempo real. Para mejorar la percepción de detalles locales, proponemos modificar el modelo de iluminación utilizando una conocida herramienta de procesado de imágenes (unsharp masking). Iluminando aquellos detalles que sobresalen de las superficies y oscureciendo las zonas profundas, se mejora el contraste local de la imagen, con lo que se consigue realzar los detalles de superficie. También se presentan diferentes técnicas para mejorar la percepción de la profundidad en Direct Volume Rendering. Concretamente, se propone modificar la iluminación teniendo en cuenta la oclusión ambiente de dos maneras diferentes: la primera utiliza los valores de profundidad en espacio imagen para calcular el factor de oclusión del entorno de cada pixel, mientras que la segunda utiliza los valores de densidad del volumen para aproximar dicha oclusión en cada vóxel. Además de estas dos técnicas, también se propone mejorar la percepción espacial y de la profundidad de ciertas estructuras mediante la generación de halos. La técnica conocida como Maximum Intensity Projection (MIP) permite visualizar los elementos de mayor intensidad del volumen, pero no aporta ningún tipo de información contextual. Para mejorar la percepción de la profundidad, proponemos una nueva técnica basada en cambiar la forma en la que se acumula la intensidad en MIP. También se describe un esquema de color para mejorar la percepción espacial de los elementos visualizados. La última contribución de la tesis es una herramienta de manipulación directa de los datos, que permite preservar la información contextual cuando se realizan cortes en el modelo de volumen. Basada en técnicas ilustrativas tradicionales, esta técnica permite al usuario estirar las estructuras visibles en las secciones de los cortes. Como resultado, las estructuras de interés se visualizan a diferentes alturas sobre la sección, lo que permite al observador percibirlas correctamente

    SOFI: A 3D simulator for the generation of underwater optical images

    No full text
    International audienceWe present an original simulator-called SOFI-for the synthetic generation of underwater optical images. The simulator architecture is flexible and relies on flow diagrams in order to allow the integration of various models for image generation which are based on the underwater optical phenomena. The objective is also to ensure real time or quasi real time performance so it takes advantage of the latest technologies, such as GPGPU, and relies on GPU programming under CUDA. Two kinds of models for image generation are presented and should be integrated in SOFI: (1) the OSOA model based on the radiative transfer theory and (2) global image modeling which describes globally how an image is deteriorated under the effects of sea water
    corecore