6,392 research outputs found

    Remote and scalable interactive high-fidelity graphics using asynchronous computation

    Get PDF
    Current computing devices span a large and varied range of computational power. Interactive high-fidelity graphics is still unachievable on many of the devices widely available to the public, such as desktops and laptops without high-end dedicated graphics cards, tablets and mobile phones. In this paper we present a scalable solution for interactive high-fidelity graphics with global illumination in the cloud. Specifically, we introduce a novel method for the asynchronous remote computation of indirect lighting that is both scalable and efficient. A lightweight client implementation merges the remotely computed indirect contribution with locally computed direct lighting for a full global illumination solution. The approach proposed in this paper applies instant radiosity methods to a precomputed point cloud representation of the scene; an equivalent structure on the client side is updated on demand, and used to reconstruct the indirect contribution. This method can be deployed on platforms of varying computational power, from tablets to high-end desktops and video game consoles. Furthermore, the same dynamic GI solution computed on the cloud can be used concurrently with multiple clients sharing a virtual environment with minimal overheads.peer-reviewe

    The Iray Light Transport Simulation and Rendering System

    Full text link
    While ray tracing has become increasingly common and path tracing is well understood by now, a major challenge lies in crafting an easy-to-use and efficient system implementing these technologies. Following a purely physically-based paradigm while still allowing for artistic workflows, the Iray light transport simulation and rendering system allows for rendering complex scenes by the push of a button and thus makes accurate light transport simulation widely available. In this document we discuss the challenges and implementation choices that follow from our primary design decisions, demonstrating that such a rendering system can be made a practical, scalable, and efficient real-world application that has been adopted by various companies across many fields and is in use by many industry professionals today

    Joint Material and Illumination Estimation from Photo Sets in the Wild

    Get PDF
    Faithful manipulation of shape, material, and illumination in 2D Internet images would greatly benefit from a reliable factorization of appearance into material (i.e., diffuse and specular) and illumination (i.e., environment maps). On the one hand, current methods that produce very high fidelity results, typically require controlled settings, expensive devices, or significant manual effort. To the other hand, methods that are automatic and work on 'in the wild' Internet images, often extract only low-frequency lighting or diffuse materials. In this work, we propose to make use of a set of photographs in order to jointly estimate the non-diffuse materials and sharp lighting in an uncontrolled setting. Our key observation is that seeing multiple instances of the same material under different illumination (i.e., environment), and different materials under the same illumination provide valuable constraints that can be exploited to yield a high-quality solution (i.e., specular materials and environment illumination) for all the observed materials and environments. Similar constraints also arise when observing multiple materials in a single environment, or a single material across multiple environments. The core of this approach is an optimization procedure that uses two neural networks that are trained on synthetic images to predict good gradients in parametric space given observation of reflected light. We evaluate our method on a range of synthetic and real examples to generate high-quality estimates, qualitatively compare our results against state-of-the-art alternatives via a user study, and demonstrate photo-consistent image manipulation that is otherwise very challenging to achieve

    From Big Data to Big Displays: High-Performance Visualization at Blue Brain

    Full text link
    Blue Brain has pushed high-performance visualization (HPV) to complement its HPC strategy since its inception in 2007. In 2011, this strategy has been accelerated to develop innovative visualization solutions through increased funding and strategic partnerships with other research institutions. We present the key elements of this HPV ecosystem, which integrates C++ visualization applications with novel collaborative display systems. We motivate how our strategy of transforming visualization engines into services enables a variety of use cases, not only for the integration with high-fidelity displays, but also to build service oriented architectures, to link into web applications and to provide remote services to Python applications.Comment: ISC 2017 Visualization at Scale worksho

    Teaching about Madrid: A Collaborative Agents-Based Distributed Learning Course

    Get PDF
    Interactive art courses require a huge amount of computational resources to be running on real time. These computational resources are even bigger if the course has been designed as a Virtual Environment with which students can interact. In this paper, we present an initiative that has been develop in a close collaboration between two Spanish Universities: Universidad Politécnica de Madrid and Universidad Rey Juan Carlos with the aim of join two previous research project: a Collaborative Awareness Model for Task-Balancing-Delivery (CAMT) in clusters and the “Teaching about Madrid” course, which provides a cultural interactive background of the capital of Spain

    Mobile graphics: SIGGRAPH Asia 2017 course

    Get PDF
    Peer ReviewedPostprint (published version
    • …
    corecore