5 research outputs found

    Hierarchical Back-Face Computation

    Full text link

    Digital Alchemy: Matter and Metamorphosis in Contemporary Digital Animation and Interface Design

    Get PDF
    The recent proliferation of special effects in Hollywood film has ushered in an era of digital transformation. Among scholars, digital technology is hailed as a revolutionary moment in the history of communication and representation. Nevertheless, media scholars and cultural historians have difficulty finding a language adequate to theorizing digital artifacts because they are not just texts to be deciphered. Rather, digital media artifacts also invite critiques about the status of reality because they resurrect ancient problems of embodiment and transcendence.In contrast to scholarly approaches to digital technology, computer engineers, interface designers, and special effects producers have invented a robust set of terms and phrases to describe the practice of digital animation. In order to address this disconnect between producers of new media and scholars of new media, I argue that the process of digital animation borrows extensively from a set of preexisting terms describing materiality that were prominent for centuries prior to the scientific revolution. Specifically, digital animators and interface designers make use of the ancient science, art, and technological craft of alchemy. Both alchemy and digital animation share several fundamental elements: both boast the power of being able to transform one material, substance, or thing into a different material, substance, or thing. Both seek to transcend the body and materiality but in the process, find that this elusive goal (realism and gold) is forever receding onto the horizon.The introduction begins with a literature review of the field of digital media studies. It identifies a gap in the field concerning disparate arguments about new media technology. On the one hand, scholars argue that new technologies like cyberspace and digital technology enable radical new forms of engagement with media on individual, social, and economic levels. At the same time that media scholars assert that our current epoch is marked by a historical rupture, many other researchers claim that new media are increasingly characterized by ancient metaphysical problems like embodiment and transcendence. In subsequent chapters I investigate this disparity

    Towards Predictive Rendering in Virtual Reality

    Get PDF
    The strive for generating predictive images, i.e., images representing radiometrically correct renditions of reality, has been a longstanding problem in computer graphics. The exactness of such images is extremely important for Virtual Reality applications like Virtual Prototyping, where users need to make decisions impacting large investments based on the simulated images. Unfortunately, generation of predictive imagery is still an unsolved problem due to manifold reasons, especially if real-time restrictions apply. First, existing scenes used for rendering are not modeled accurately enough to create predictive images. Second, even with huge computational efforts existing rendering algorithms are not able to produce radiometrically correct images. Third, current display devices need to convert rendered images into some low-dimensional color space, which prohibits display of radiometrically correct images. Overcoming these limitations is the focus of current state-of-the-art research. This thesis also contributes to this task. First, it briefly introduces the necessary background and identifies the steps required for real-time predictive image generation. Then, existing techniques targeting these steps are presented and their limitations are pointed out. To solve some of the remaining problems, novel techniques are proposed. They cover various steps in the predictive image generation process, ranging from accurate scene modeling over efficient data representation to high-quality, real-time rendering. A special focus of this thesis lays on real-time generation of predictive images using bidirectional texture functions (BTFs), i.e., very accurate representations for spatially varying surface materials. The techniques proposed by this thesis enable efficient handling of BTFs by compressing the huge amount of data contained in this material representation, applying them to geometric surfaces using texture and BTF synthesis techniques, and rendering BTF covered objects in real-time. Further approaches proposed in this thesis target inclusion of real-time global illumination effects or more efficient rendering using novel level-of-detail representations for geometric objects. Finally, this thesis assesses the rendering quality achievable with BTF materials, indicating a significant increase in realism but also confirming the remainder of problems to be solved to achieve truly predictive image generation

    Interactive Display of Large-Scale NURBS Models \Lambda Subodh Kumar Dinesh Manocha Anselmo Lastra

    No full text
    Current graphics systems have reached the capability of rendering millions of transformed, shaded and z-buffered polygons per second [3, 14]. However in many applications involving CAD/CAM, virtual reality, animation and visualization the object models are described in terms of non-uniform rational B-spline (NURBS) surfaces. This class includes B'ezier surfaces and other rational parametric surfaces like tensor product and triangular patches. Large scale models consisting of thousands of such surfaces are commonly used to represent shapes of automobiles, submarines, airplanes, building architectures, sculptured models, mechanical parts and in applications involving surface fitting over scattered data or surface reconstruction. Current renderers of sculptured models on commercial graphics systems, while faster than ever before, are not able to render them in real time for applications involving virtual worlds, walkthroughs and other immersive technologies
    corecore