166,260 research outputs found

    Towards a Scalable Architecture for Real-Time Volume Rendering

    Get PDF
    In this paper we present our research eff orts towards a scalable volume rendering architecture for the real-time visualization of dynamically changing high-resolution datasets. Using a linearly skewed memory interleaving we were able to develop a parallel data ow model that leads to local, fixed-bandwidth interconnections between processing elements. This parallel dataflow model diff ers from previous work in that it requires no global communication of data except at the pixel level. Using this data ow model we are developing Cube-4, an architecture that is scalable to very high performances and allows for modular and extensible hardware implementations.Engineering and Applied Science

    VIRIM, A Real-Time Volume Rendering System for Medicine

    Full text link
    VIRIM, a real-time direct volume rendering system is presented. The system is freely programmable and supports models like a-compositing, front-to-back (back-to-front) techniques, and the slab method. The hardware system is divided into two units, a geometry unit and a raycast unit. The geometry unit performs resampling and gradient estimation and is mapped directly into hardware. It supports different resampling filters in order to reduce resampling artifacts. The raycast unit consists of 16 digital signal processors that perform the programmable ray- casting. The software of VIRIM is layered and provides manipulation tools for the data during real-time visualization like arbitrary gray-value mapping and setting the region of interest. The system is under test and will be available as prototype in 1995

    Three architectures for volume rendering

    Get PDF
    Volume rendering is a key technique in scientific visualization that lends itself to significant exploitable parallelism. The high computational demands of real-time volume rendering and continued technological advances in the area of VLSI give impetus to the development of special-purpose volume rendering architectures. This paper presents and characterizes three recently developed volume rendering engines which are based on the ray-casting method. A taxonomy of the algorithmic variants of ray-casting and details of each ray-casting architecture are discussed. The paper then compares the machine features and provides an outlook on future developments in the area of volume rendering hardware

    Real-time volume rendering and tractography visualization on the web

    Get PDF
    In the field of computer graphics, Volume Rendering techniques allow the visualization of 3D datasets, and specifically, Volume Ray-Casting renders images from volumetric datasets, typically used in some scientific areas, such as medical imaging -- This article aims to describe the development of a combined visualization of tractography and volume rendering of brain T1 MRI images in an integrated way -- An innovative web viewer for interactive visualization of neuro-imaging data has been developed based on WebGL -- This recently developed standard enables the clients to use the web viewer on a wide range of devices, with the only requirement of a compliant web-browser -- As the majority of the rendering tasks take place in the client machine, the effect of bottlenecks and server overloading are minimized -- The web application presented is able to compete with desktop tools, even supporting high graphical demands and facing challenges regarding performance and scalability -- The developed software modules are available as open source code and include MRI volume data and tractography generated by the Diffusion Toolkit, and connectivity data from the Connectome Mapping Toolkit -- Our contribution for the Volume Web Viewer implements early ray termination step according to the tractography depthmap, combining volume images and estimated white matter fibers -- Furthermore, the depthmap system extension can be used for visualization of other types of data, where geometric and volume elements are displayed simultaneousl

    Volumetric real-time particle-based representation of large unstructured tetrahedral polygon meshes

    No full text
    In this paper we propose a particle-based volume rendering approach for unstructured, three-dimensional, tetrahedral polygon meshes. We stochastically generate millions of particles per second and project them on the screen in real-time. In contrast to previous rendering techniques of tetrahedral volume meshes, our method does not need a prior depth sorting of geometry. Instead, the rendered image is generated by choosing particles closest to the camera. Furthermore, we use spatial superimposing. Each pixel is constructed from multiple subpixels. This approach not only increases projection accuracy, but allows also a combination of subpixels into one superpixel that creates the well-known translucency effect of volume rendering. We show that our method is fast enough for the visualization of unstructured three-dimensional grids with hard real-time constraints and that it scales well for a high number of particles

    Volume ray casting techniques and applications using general purpose computations on graphics processing units

    Get PDF
    Traditional 3D computer graphics focus on rendering the exterior of objects. Volume rendering is a technique used to visualize information corresponding to the interior of an object, commonly used in medical imaging and other fields. Visualization of such data may be accomplished by ray casting; an embarrassingly parallel algorithm also commonly used in ray tracing. There has been growing interest in performing general purpose computations on graphics processing units (GPGPU), which are capable exploiting parallel applications and yielding far greater performance than sequential implementations on CPUs. Modern GPUs allow for rapid acceleration of volume rendering applications, offering affordable high performance visualization systems. This thesis explores volume ray casting performance and visual quality enhancements using the NVIDIA CUDA platform, and demonstrates how high quality volume renderings can be produced with interactive and real time frame rates on modern commodity graphics hardware. A number of techniques are employed in this effort, including early ray termination, super sampling and texture filtering. In a performance comparison of a sequential versus CUDA implementation on high-end hardware, the latter is capable of rendering 60 frames per second with an impressive price-performance ratio heavily favoring GPUs. A number of unique volume rendering applications are explored including multiple volume rendering capable of arbitrary placement and rigid volume registration, hypertexturing and stereoscopic anaglyphs, each greatly enhanced by the real time interaction of volume data. The techniques and applications discussed in this thesis may prove to be invaluable tools in fields such as medical and molecular imaging, flow and scientific visualization, engineering drawing and many others

    Efficient light scattering through thin semi-transparent objects

    Get PDF
    This paper concerns real-time rendering of thin semi-transparent objects. An object in this category could be a piece of cloth, eg. a curtain. Semi-transparent objects are visualized most correctly using volume rendering techniques. In general such techniques are, however, intractable for real-time applications. Surface rendering is more efficient, but also inadequate since semi-transparent objects should have a different appearance depending on whether they are front-lit or back-lit. The back-lit side of a curtain, for example, often seems quite transparent while the front-lit side seems brighter and almost opaque. To capture such visual effects in the standard rendering pipeline, Blinn [1982] proposed an efficient local illumination model based on radiative transfer theory. He assumed media of low density, hence, his equations can render media such as clouds, smoke, and dusty surfaces. Our observation is that Chandrasekhar [1960] has derived the same equations from a different set of assumptions. This alternative derivation makes the theory useful for realistic real-time rendering of dense, but thin, semitransparent objects such as cloth. We demonstrate that application of the theory in this new area gives far better results than what is obtainable with a traditional real-time rendering scheme using a constant factor for alpha blending
    corecore