550 research outputs found
Recommended from our members
Acquisition and Rendering of Transparent and Refractive Objects
This paper introduces a new image-based approach to capturing and modeling highly specular, transparent, or translucent objects. We have built a system for automatically acquiring high quality graphical models of objects that are extremely difficult to scan with traditional 3D scanners. The system consists of turntables, a set of cameras and lights, and monitors to project colored backdrops. We use multi-background matting techniques to acquire alpha and environment mattes of the object from multiple viewpoints. Using the alpha mattes we reconstruct an approximate 3D shape of the object. We use the environment mattes to compute a high-resolution surface reflectance field. We also acquire a low-resolution surface reflectance field using the overhead array of lights. Both surface reflectance fields are used to relight the objects and to place them into arbitrary environments. Our system is the first to acquire and render transparent and translucent 3D objects, such as a glass of beer, from arbitrary viewpoints under novel illumination.Engineering and Applied Science
Recommended from our members
Object Space EWA Surface Splatting: A Hardware Accelerated Approach to High Quality Point Rendering
Elliptical weighted average (EWA) surface splatting is a technique for high quality rendering of point-sampled 3D objects. EWA surface splatting renders water-tight surfaces of complex point models with high quality, anisotropic texture filtering. In this paper we introduce a new multi-pass approach to perform EWA surface splatting on modern PC graphics hardware, called object space EWA splatting. We derive an object space formulation of the EWA filter, which is amenable for acceleration by conventional triangle-based graphics hardware. We describe how to implement the object space EWA filter using a two pass rendering algorithm. In the first rendering pass, visibility splatting is performed by shifting opaque surfel polygons backward along the viewing rays, while in the second rendering pass view-dependent EWA prefiltering is performed by deforming texture mapped surfel polygons. We use texture mapping and alpha blending to facilitate the splatting process. We implement our algorithm using programmable vertex and pixel shaders, fully exploiting the capabilities of today’s graphics processing units (GPUs). Our implementation renders up to 3 million points per second on recent PC graphics hardware, an order of magnitude more than a pure software implementation of screen space EWA surface splatting.Engineering and Applied Science
Procedural Generation and Rendering of Realistic, Navigable Forest Environments: An Open-Source Tool
Simulation of forest environments has applications from entertainment and art
creation to commercial and scientific modelling. Due to the unique features and
lighting in forests, a forest-specific simulator is desirable, however many
current forest simulators are proprietary or highly tailored to a particular
application. Here we review several areas of procedural generation and
rendering specific to forest generation, and utilise this to create a
generalised, open-source tool for generating and rendering interactive,
realistic forest scenes. The system uses specialised L-systems to generate
trees which are distributed using an ecosystem simulation algorithm. The
resulting scene is rendered using a deferred rendering pipeline, a Blinn-Phong
lighting model with real-time leaf transparency and post-processing lighting
effects. The result is a system that achieves a balance between high natural
realism and visual appeal, suitable for tasks including training computer
vision algorithms for autonomous robots and visual media generation.Comment: 14 pages, 11 figures. Submitted to Computer Graphics Forum (CGF). The
application and supporting configuration files can be found at
https://github.com/callumnewlands/ForestGenerato
Interactive global illumination on the CPU
Computing realistic physically-based global illumination in real-time remains one
of the major goals in the fields of rendering and visualisation; one that has not
yet been achieved due to its inherent computational complexity. This thesis focuses
on CPU-based interactive global illumination approaches with an aim to
develop generalisable hardware-agnostic algorithms. Interactive ray tracing is reliant
on spatial and cache coherency to achieve interactive rates which conflicts
with needs of global illumination solutions which require a large number of incoherent
secondary rays to be computed. Methods that reduce the total number of
rays that need to be processed, such as Selective rendering, were investigated to
determine how best they can be utilised.
The impact that selective rendering has on interactive ray tracing was analysed
and quantified and two novel global illumination algorithms were developed,
with the structured methodology used presented as a framework. Adaptive Inter-
leaved Sampling, is a generalisable approach that combines interleaved sampling
with an adaptive approach, which uses efficient component-specific adaptive guidance
methods to drive the computation. Results of up to 11 frames per second
were demonstrated for multiple components including participating media. Temporal Instant Caching, is a caching scheme for accelerating the computation of
diffuse interreflections to interactive rates. This approach achieved frame rates
exceeding 9 frames per second for the majority of scenes. Validation of the results
for both approaches showed little perceptual difference when comparing
against a gold-standard path-traced image. Further research into caching led to
the development of a new wait-free data access control mechanism for sharing the
irradiance cache among multiple rendering threads on a shared memory parallel
system. By not serialising accesses to the shared data structure the irradiance
values were shared among all the threads without any overhead or contention,
when reading and writing simultaneously. This new approach achieved efficiencies
between 77% and 92% for 8 threads when calculating static images and animations.
This work demonstrates that, due to the
flexibility of the CPU, CPU-based
algorithms remain a valid and competitive choice for achieving global illumination
interactively, and an alternative to the generally brute-force GPU-centric
algorithms
Painterly rendering techniques: A state-of-the-art review of current approaches
In this publication we will look at the different methods presented over the past few decades which attempt to recreate digital paintings. While previous surveys concentrate on the broader subject of non-photorealistic rendering, the focus of this paper is firmly placed on painterly rendering techniques. We compare different methods used to produce different output painting styles such as abstract, colour pencil, watercolour, oriental, oil and pastel. Whereas some methods demand a high level of interaction using a skilled artist, others require simple parameters provided by a user with little or no artistic experience. Many methods attempt to provide more automation with the use of varying forms of reference data. This reference data can range from still photographs, video, 3D polygonal meshes or even 3D point clouds. The techniques presented here endeavour to provide tools and styles that are not traditionally available to an artist. Copyright © 2012 John Wiley & Sons, Ltd
A Word about the Weather: depiction and visualisation
‘Finally, a word about the Weather.’ With these words, the BBC Trust began an unusual paragraph in their 2007 report on impartiality (BBC Trust 2007:53). The advent of a new three-dimensional visualisation of the weather in May 2005 had prompted immediate criticism. The images suggested northern Scotland was on the periphery, while south-east England was in the forefront. In its own odd way this controversy touches on an important issue in visualisation and depiction, and on the relationship between those two things.
This was a paper for the AHRC IT Methods Workshop ‘From Abstract Data Mapping to 3D Photorealism.’ The workshop title could be taken to imply that abstract data mapping (visualisation) and photorealism (an extreme form of depiction) are as far apart as can be, perhaps even that they are opposites. By contrast, this paper explores some aspects they have in common. Boyd Davis argues that it is productive to think of depiction as a form of visualisation, and to think of it as therefore subject to many of the same design considerations
- …