1,520 research outputs found
A survey of real-time crowd rendering
In this survey we review, classify and compare existing approaches for real-time crowd rendering. We first overview character animation techniques, as they are highly tied to crowd rendering performance, and then we analyze the state of the art in crowd rendering. We discuss different representations for level-of-detail (LoD) rendering of animated characters, including polygon-based, point-based, and image-based techniques, and review different criteria for runtime LoD selection. Besides LoD approaches, we review classic acceleration schemes, such as frustum culling and occlusion culling, and describe how they can be adapted to handle crowds of animated characters. We also discuss specific acceleration techniques for crowd rendering, such as primitive pseudo-instancing, palette skinning, and dynamic key-pose caching, which benefit from current graphics hardware. We also address other factors affecting performance and realism of crowds such as lighting, shadowing, clothing and variability. Finally we provide an exhaustive comparison of the most relevant approaches in the field.Peer ReviewedPostprint (author's final draft
An Intestinal Surgery Simulator: Real-Time Collision Processing and Visualization
International audienceThis research work is aimed towards the development of a VR-based trainer for colon cancer removal. It enables the surgeons to interactively view and manipulate the concerned virtual organs as during a real surgery. First, we present a method for animating the small intestine and the mesentery (the tissue that connects it to the main vessels) in real-time, thus enabling user-interaction through virtual surgical tools during the simulation. We present a stochastic approach for fast collision detection in highly deformable, self-colliding objects. A simple and efficient response to collisions is also introduced in order to reduce the overall animation complexity. Secondly, we describe a new method based on generalized cylinders for fast rendering of the intestine. An efficient curvature detection method, along with an adaptive sampling algorithm is presented. This approach, while providing improved tessellation without the classical self-intersection problem, also allows for high-performance rendering, thanks to the new 3D skinning feature available in recent GPUs. The rendering algorithm is also designed to ensure a guaranteed frame rate. Finally, we present the quantitative results of the simulations and describe the qualitative feedback obtained from the surgeons
Sketching-out virtual humans: A smart interface for human modelling and animation
In this paper, we present a fast and intuitive interface for sketching out
3D virtual humans and animation. The user draws stick figure key frames first and
chooses one for “fleshing-out” with freehand body contours. The system
automatically constructs a plausible 3D skin surface from the rendered figure, and
maps it onto the posed stick figures to produce the 3D character animation. A
“creative model-based method” is developed, which performs a human perception
process to generate 3D human bodies of various body sizes, shapes and fat
distributions. In this approach, an anatomical 3D generic model has been created with
three distinct layers: skeleton, fat tissue, and skin. It can be transformed sequentially
through rigid morphing, fatness morphing, and surface fitting to match the original
2D sketch. An auto-beautification function is also offered to regularise the 3D
asymmetrical bodies from users’ imperfect figure sketches. Our current system
delivers character animation in various forms, including articulated figure animation,
3D mesh model animation, 2D contour figure animation, and even 2D NPR animation
with personalised drawing styles. The system has been formally tested by various
users on Tablet PC. After minimal training, even a beginner can create vivid virtual
humans and animate them within minutes
Multifarious Hierarchies of Mechanical Models for Artist Assigned Levels-of-Detail
International audienceWe present a new framework for artist driven level of detail in solid simulations. Simulated objects are simultaneously embedded in several, separately designed deformation models with their own independent degrees of freedom. The models are ordered to apply their deformations hierarchically, and we enforce the uniqueness of the dynamics solutions using a novel kinetic filtering operator designed to ensure that each child only adds detail motion to its parent without introducing redundancies. This new approach allows artists to easily add fine-scale details without introducing unnecessary degrees-of-freedom to the simulation or resorting to complex geometric operations like anisotropic volume meshing. We illustrate the utility of our approach with several detail enriched simulation examples
Embedded Implicit Stand-ins for Animated Meshes: a Case of Hybrid Modelling
In this paper we address shape modelling problems, encountered in computer animation and computer games development that are difficult to solve just using polygonal meshes. Our approach is based on a hybrid modelling concept that combines polygonal meshes with implicit surfaces. A hybrid model consists of an animated polygonal mesh and an approximation of this mesh by a convolution surface stand-in that is embedded within it or is attached to it. The motions of both objects are synchronised using a rigging skeleton. This approach is used to model the interaction between an animated mesh object and a viscoelastic substance, normally modelled in implicit form. The adhesive behaviour of the viscous object is modelled using geometric blending operations on the corresponding implicit surfaces. Another application of this approach is the creation of metamorphosing implicit surface parts that are attached to an animated mesh. A prototype implementation of the proposed approach and several examples of modelling and animation with near real-time preview times are presented
Virtual humans: thirty years of research, what next?
In this paper, we present research results and future challenges in creating realistic and believable Virtual Humans. To realize these modeling goals, real-time realistic representation is essential, but we also need interactive and perceptive Virtual Humans to populate the Virtual Worlds. Three levels of modeling should be considered to create these believable Virtual Humans: 1) realistic appearance modeling, 2) realistic, smooth and flexible motion modeling, and 3) realistic high-level behaviors modeling. At first, the issues of creating virtual humans with better skeleton and realistic deformable bodies are illustrated. To give a level of believable behavior, challenges are laid on generating on the fly flexible motion and complex behaviours of Virtual Humans inside their environments using a realistic perception of the environment. Interactivity and group behaviours are also important parameters to create believable Virtual Humans which have challenges in creating believable relationship between real and virtual humans based on emotion and personality, and simulating realistic and believable behaviors of groups and crowds. Finally, issues in generating realistic virtual clothed and haired people are presente
Recommended from our members
View-dependent adaptive cloth simulation
This paper describes a method for view-dependent cloth simulation using dynamically adaptive mesh refinement and coarsening. Given a prescribed camera motion, the method adjusts the criteria controlling refinement to account for visibility and apparent size in the camera's view. Objectionable dynamic artifacts are avoided by anticipative refinement and smoothed coarsening. This approach preserves the appearance of detailed cloth throughout the animation while avoiding the wasted effort of simulating details that would not be discernible to the viewer. The computational savings realized by this method increase as scene complexity grows, producing a 2× speed-up for a single character and more than 4× for a small group
- …