6,392 research outputs found
Remote and scalable interactive high-fidelity graphics using asynchronous computation
Current computing devices span a large and varied range of computational power. Interactive high-fidelity graphics is still unachievable on many of the devices widely available to the public, such as desktops and laptops without high-end dedicated graphics cards, tablets and mobile phones. In this paper we present a scalable solution for interactive high-fidelity graphics with global illumination in the cloud. Specifically, we introduce a novel method for the asynchronous remote computation of indirect lighting that is both scalable and efficient. A lightweight client implementation merges the remotely computed indirect contribution with locally computed direct lighting for a full global illumination solution. The approach proposed in this paper applies instant radiosity methods to a precomputed point cloud representation of the scene; an equivalent structure on the client side is updated on demand, and used to reconstruct the indirect contribution. This method can be deployed on platforms of varying computational power, from tablets to high-end desktops and video game consoles. Furthermore, the same dynamic GI solution computed on the cloud can be used concurrently with multiple clients sharing a virtual environment with minimal overheads.peer-reviewe
The Iray Light Transport Simulation and Rendering System
While ray tracing has become increasingly common and path tracing is well
understood by now, a major challenge lies in crafting an easy-to-use and
efficient system implementing these technologies. Following a purely
physically-based paradigm while still allowing for artistic workflows, the Iray
light transport simulation and rendering system allows for rendering complex
scenes by the push of a button and thus makes accurate light transport
simulation widely available. In this document we discuss the challenges and
implementation choices that follow from our primary design decisions,
demonstrating that such a rendering system can be made a practical, scalable,
and efficient real-world application that has been adopted by various companies
across many fields and is in use by many industry professionals today
Joint Material and Illumination Estimation from Photo Sets in the Wild
Faithful manipulation of shape, material, and illumination in 2D Internet
images would greatly benefit from a reliable factorization of appearance into
material (i.e., diffuse and specular) and illumination (i.e., environment
maps). On the one hand, current methods that produce very high fidelity
results, typically require controlled settings, expensive devices, or
significant manual effort. To the other hand, methods that are automatic and
work on 'in the wild' Internet images, often extract only low-frequency
lighting or diffuse materials. In this work, we propose to make use of a set of
photographs in order to jointly estimate the non-diffuse materials and sharp
lighting in an uncontrolled setting. Our key observation is that seeing
multiple instances of the same material under different illumination (i.e.,
environment), and different materials under the same illumination provide
valuable constraints that can be exploited to yield a high-quality solution
(i.e., specular materials and environment illumination) for all the observed
materials and environments. Similar constraints also arise when observing
multiple materials in a single environment, or a single material across
multiple environments. The core of this approach is an optimization procedure
that uses two neural networks that are trained on synthetic images to predict
good gradients in parametric space given observation of reflected light. We
evaluate our method on a range of synthetic and real examples to generate
high-quality estimates, qualitatively compare our results against
state-of-the-art alternatives via a user study, and demonstrate
photo-consistent image manipulation that is otherwise very challenging to
achieve
From Big Data to Big Displays: High-Performance Visualization at Blue Brain
Blue Brain has pushed high-performance visualization (HPV) to complement its
HPC strategy since its inception in 2007. In 2011, this strategy has been
accelerated to develop innovative visualization solutions through increased
funding and strategic partnerships with other research institutions.
We present the key elements of this HPV ecosystem, which integrates C++
visualization applications with novel collaborative display systems. We
motivate how our strategy of transforming visualization engines into services
enables a variety of use cases, not only for the integration with high-fidelity
displays, but also to build service oriented architectures, to link into web
applications and to provide remote services to Python applications.Comment: ISC 2017 Visualization at Scale worksho
Teaching about Madrid: A Collaborative Agents-Based Distributed Learning Course
Interactive art courses require a huge amount of computational resources to be running on real time. These computational resources are even bigger if the course has been designed as a Virtual Environment with which students can interact. In this paper, we present an initiative that has been develop in a close collaboration between two Spanish Universities: Universidad Politécnica de Madrid and Universidad Rey Juan Carlos with the aim of join two previous research project: a Collaborative Awareness Model for Task-Balancing-Delivery (CAMT) in clusters and the “Teaching about Madrid” course, which provides a cultural interactive background of the capital of Spain
Mobile graphics: SIGGRAPH Asia 2017 course
Peer ReviewedPostprint (published version
- …