2,333 research outputs found
Creating Simplified 3D Models with High Quality Textures
This paper presents an extension to the KinectFusion algorithm which allows
creating simplified 3D models with high quality RGB textures. This is achieved
through (i) creating model textures using images from an HD RGB camera that is
calibrated with Kinect depth camera, (ii) using a modified scheme to update
model textures in an asymmetrical colour volume that contains a higher number
of voxels than that of the geometry volume, (iii) simplifying dense polygon
mesh model using quadric-based mesh decimation algorithm, and (iv) creating and
mapping 2D textures to every polygon in the output 3D model. The proposed
method is implemented in real-time by means of GPU parallel processing.
Visualization via ray casting of both geometry and colour volumes provides
users with a real-time feedback of the currently scanned 3D model. Experimental
results show that the proposed method is capable of keeping the model texture
quality even for a heavily decimated model and that, when reconstructing small
objects, photorealistic RGB textures can still be reconstructed.Comment: 2015 International Conference on Digital Image Computing: Techniques
and Applications (DICTA), Page 1 -
RTSDF: Generating Signed Distance Fields in Real Time for Soft Shadow Rendering
Signed Distance Fields (SDFs) for surface representation are commonly
generated offline and subsequently loaded into interactive applications like
games. Since they are not updated every frame, they only provide a rigid
surface representation. While there are methods to generate them quickly on
GPU, the efficiency of these approaches is limited at high resolutions. This
paper showcases a novel technique that combines jump flooding and ray tracing
to generate approximate SDFs in real-time for soft shadow approximation,
achieving prominent shadow penumbras while maintaining interactive frame rates
3D rekonstrukce na iOS
This bachelor thesis describes implementation of a real-time RGBD-based 3D reconstruction pipeline suited for Apple’s iPhone X with the TrueDepth camera. First, an overview of common approaches to the reconstruction problem is made, followed by a description of the underlying algorithms and techniques used in the thesis. Finally, the implementation details of the application pipeline are presented with performance overview of the implemented application.Tato bakalářská práce popisuje implementaci řetězce pro 3D rekonstrukci z RGBD snímků v reálném čase, určené pro Apple iPhone X s TrueDepth kamerou. Nejdříve je podán přehled běžných přístupů k rekonstrukci, následován popisem algoritmů a technik použitých v této práci. Nakonec jsou popsány implementační detaily zvoleného rekonstrukčního řetězce spolu s popisem výkonnosti implementované aplikace.460 - Katedra informatikyvýborn
Volumetric Isosurface Rendering with Deep Learning-Based Super-Resolution
Rendering an accurate image of an isosurface in a volumetric field typically
requires large numbers of data samples. Reducing the number of required samples
lies at the core of research in volume rendering. With the advent of deep
learning networks, a number of architectures have been proposed recently to
infer missing samples in multi-dimensional fields, for applications such as
image super-resolution and scan completion. In this paper, we investigate the
use of such architectures for learning the upscaling of a low-resolution
sampling of an isosurface to a higher resolution, with high fidelity
reconstruction of spatial detail and shading. We introduce a fully
convolutional neural network, to learn a latent representation generating a
smooth, edge-aware normal field and ambient occlusions from a low-resolution
normal and depth field. By adding a frame-to-frame motion loss into the
learning stage, the upscaling can consider temporal variations and achieves
improved frame-to-frame coherence. We demonstrate the quality of the network
for isosurfaces which were never seen during training, and discuss remote and
in-situ visualization as well as focus+context visualization as potential
application
Highly parallel Monte-Carlo simulations of the acousto-optic effect in heterogeneous turbid media
The development of a highly parallel simulation of the acousto-optic effect is detailed. The simulation supports optically heterogeneous simulation domains under insonification by arbitrary monochromatic ultrasound fields. An adjoint method for acousto-optics is proposed to permit point-source/point-detector simulations. The flexibility and efficiency of this simulation code is demonstrated in the development of spatial absorption sensitivity maps which are in broad agreement with current experimental investigations. The simulation code has the potential to provide guidance in the feasibility and optimization of future studies of the acousto-optic technique, and its speed may permit its use as part of an iterative inversion model
Enhanced perception in volume visualization
Due to the nature of scientic data sets, the generation of convenient visualizations may be a difficult task, but crucial to correctly convey the relevant information of the data. When working with complex volume models, such as the anatomical ones, it is important to provide accurate representations, since a misinterpretation can lead to serious mistakes while diagnosing a disease or planning surgery. In these cases, enhancing the perception of the features of interest usually helps to properly understand the data.
Throughout years, researchers have focused on different methods to improve the visualization of volume data sets. For instance, the definition of good transfer functions is a key issue in Volume Visualization, since transfer functions determine how materials are classified. Other approaches are based on simulating realistic illumination models to enhance the spatial perception, or using illustrative effects to provide the level of abstraction needed to correctly interpret the data.
This thesis contributes with new approaches to enhance the visual and spatial perception in Volume Visualization. Thanks to the new computing capabilities of modern graphics hardware, the proposed algorithms are capable of modifying the illumination model and simulating illustrative motifs in real time.
In order to enhance local details, which are useful to better perceive the shape and the surfaces of the volume, our first contribution is an algorithm that employs a common sharpening operator to modify the lighting applied. As a result, the overall contrast of the visualization is enhanced by brightening the salient features and darkening the deeper regions of the volume model.
The enhancement of depth perception in Direct Volume Rendering is also covered in the thesis. To do this, we propose two algorithms to simulate ambient occlusion: a screen-space technique based on using depth information to estimate the amount of light occluded, and a view-independent method that uses the density values of the data set to estimate the occlusion. Additionally, depth perception is also enhanced by adding halos around the structures of interest.
Maximum Intensity Projection images provide a good understanding of the high intensity features of the data, but lack any contextual information. In order to enhance the depth perception in such a case, we present a novel technique based on changing how intensity is accumulated. Furthermore, the perception of the spatial arrangement of the displayed structures is also enhanced by adding certain colour cues.
The last contribution is a new manipulation tool designed for adding contextual information when cutting the volume. Based on traditional illustrative effects, this method allows the user to directly extrude structures from the cross-section of the cut. As a result, the clipped structures are displayed at different heights, preserving the information needed to correctly perceive them.Debido a la naturaleza de los datos científicos, visualizarlos correctamente puede ser una tarea complicada, pero crucial para interpretarlos de forma adecuada. Cuando se trabaja con modelos de volumen complejos, como es el caso de los modelos anatómicos, es importante generar imágenes precisas, ya que una mala interpretación de las mismas puede producir errores graves en el diagnóstico de enfermedades o en la planificación de operaciones quirúrgicas. En estos casos, mejorar la percepción de las zonas de interés, facilita la comprensión de la información inherente a los datos.
Durante décadas, los investigadores se han centrado en el desarrollo de técnicas para mejorar la visualización de datos volumétricos. Por ejemplo,
los métodos que permiten definir buenas funciones de transferencia son clave, ya que éstas determinan cómo se clasifican los materiales. Otros ejemplos son las técnicas que simulan modelos de iluminación realista, que permiten percibir mejor la distribución espacial de los elementos del volumen, o bien
los que imitan efectos ilustrativos, que proporcionan el nivel de abstracción necesario para interpretar correctamente los datos.
El trabajo presentado en esta tesis se centra en mejorar la percepción de los elementos del volumen, ya sea modificando el modelo de iluminación aplicado en la visualización, o simulando efectos ilustrativos. Aprovechando la capacidad de cálculo de los nuevos procesadores gráficos, se describen un conjunto de algoritmos que permiten obtener los resultados en tiempo real.
Para mejorar la percepción de detalles locales, proponemos modificar el modelo de iluminación utilizando una conocida herramienta de procesado de imágenes (unsharp masking). Iluminando aquellos detalles que sobresalen de las superficies y oscureciendo las zonas profundas, se mejora el contraste local de la imagen, con lo que se consigue realzar los detalles de superficie.
También se presentan diferentes técnicas para mejorar la percepción de la profundidad en Direct Volume Rendering. Concretamente, se propone modificar la iluminación teniendo en cuenta la oclusión ambiente de dos maneras diferentes: la primera utiliza los valores de profundidad en espacio imagen para calcular el factor de oclusión del entorno de cada pixel, mientras que la segunda utiliza los valores de densidad del volumen para aproximar dicha oclusión en cada vóxel. Además de estas dos técnicas, también se propone mejorar la percepción espacial y de la profundidad de ciertas estructuras mediante la generación de halos.
La técnica conocida como Maximum Intensity Projection (MIP) permite visualizar los elementos de mayor intensidad del volumen, pero no aporta
ningún tipo de información contextual. Para mejorar la percepción de la profundidad, proponemos una nueva técnica basada en cambiar la forma en
la que se acumula la intensidad en MIP. También se describe un esquema de color para mejorar la percepción espacial de los elementos visualizados.
La última contribución de la tesis es una herramienta de manipulación directa de los datos, que permite preservar la información contextual cuando se realizan cortes en el modelo de volumen. Basada en técnicas ilustrativas tradicionales, esta técnica permite al usuario estirar las estructuras visibles en las secciones de los cortes. Como resultado, las estructuras de interés se visualizan a diferentes alturas sobre la sección, lo que permite al observador percibirlas correctamente
- …