806 research outputs found

    A tutorial on motion capture driven character animation

    Get PDF
    Motion capture (MoCap) is an increasingly important technique to create realistic human motion for animation. However MoCap data are noisy, the resulting animation is often inaccurate and unrealistic without elaborate manual processing of the data. In this paper, we will discuss practical issues for MoCap driven character animation, particularly when using commercial toolkits. We highlight open topics in this field for future research. MoCap animations created in this project will be demonstrated at the conference

    Prominence eruption initiated by helical kink-instability of an embedded flux rope

    Full text link
    We study the triggering mechanism of a limb-prominence eruption and the associated coronal mass ejection near AR 12342 using SDO and LASCO/SOHO observations. The prominence is seen with an embedded flux thread (FT) at one end and bifurcates from the middle to a different footpoint location. The morphological evolution of the FT is similar to an unstable flux rope (FR), which we regard as prominence embedded FR. The FR twist exceeds the critical value. In addition, the morphology of the prominence plasma in 304\AA~images marks the helical nature of the magnetic skeleton with a total of 2.96 turns along arc length. The potential field extrapolation model indicates that the critical height of the background magnetic field gradient falls within the inner corona (105Mm) consistent with the extent of coronal plasma loops. These results suggest that the helical kink instability in the embedded FR caused the slow rise of the prominence to a height of the torus instability domain. Moreover, the differential emission measure analysis unveils heating of the prominence plasma to coronal temperatures during eruption, suggesting a reconnection-related heating underneath the upward rising embedded FR. The prominence starts with a slow rise motion of 10km/s, followed by fast and slow acceleration phases having an average acceleration of 28.9m/s228.9m/s^2, 2.4m/s22.4m/s^2 in C2, C3 field of view respectively. As predicted by previous numerical simulations, the observed synchronous kinematic profiles of the CME leading edge and the core supports the involved FR instability in the prominence initiation.Comment: Accepted in ApJ, 13 pages, 9 figure

    Survey of Finite Element Method-Based Real-Time Simulations

    Get PDF
    The finite element method (FEM) has deservedly gained the reputation of the most powerful, highly efficient, and versatile numerical method in the field of structural analysis. Though typical application of FE programs implies the so-called “off-line” computations, the rapid pace of hardware development over the past couple of decades was the major impetus for numerous researchers to consider the possibility of real-time simulation based on FE models. Limitations of available hardware components in various phases of developments demanded remarkable innovativeness in the quest for suitable solutions to the challenge. Different approaches have been proposed depending on the demands of the specific field of application. Though it is still a relatively young field of work in global terms, an immense amount of work has already been done calling for a representative survey. This paper aims to provide such a survey, which of course cannot be exhaustive

    HIGH QUALITY HUMAN 3D BODY MODELING, TRACKING AND APPLICATION

    Get PDF
    Geometric reconstruction of dynamic objects is a fundamental task of computer vision and graphics, and modeling human body of high fidelity is considered to be a core of this problem. Traditional human shape and motion capture techniques require an array of surrounding cameras or subjects wear reflective markers, resulting in a limitation of working space and portability. In this dissertation, a complete process is designed from geometric modeling detailed 3D human full body and capturing shape dynamics over time using a flexible setup to guiding clothes/person re-targeting with such data-driven models. As the mechanical movement of human body can be considered as an articulate motion, which is easy to guide the skin animation but has difficulties in the reverse process to find parameters from images without manual intervention, we present a novel parametric model, GMM-BlendSCAPE, jointly taking both linear skinning model and the prior art of BlendSCAPE (Blend Shape Completion and Animation for PEople) into consideration and develop a Gaussian Mixture Model (GMM) to infer both body shape and pose from incomplete observations. We show the increased accuracy of joints and skin surface estimation using our model compared to the skeleton based motion tracking. To model the detailed body, we start with capturing high-quality partial 3D scans by using a single-view commercial depth camera. Based on GMM-BlendSCAPE, we can then reconstruct multiple complete static models of large pose difference via our novel non-rigid registration algorithm. With vertex correspondences established, these models can be further converted into a personalized drivable template and used for robust pose tracking in a similar GMM framework. Moreover, we design a general purpose real-time non-rigid deformation algorithm to accelerate this registration. Last but not least, we demonstrate a novel virtual clothes try-on application based on our personalized model utilizing both image and depth cues to synthesize and re-target clothes for single-view videos of different people

    Matrix-based Parameterizations of Skeletal Animated Appearance

    Full text link
    Alors que le rendu réaliste gagne de l’ampleur dans l’industrie, les techniques à la fois photoréalistes et basées sur la physique, complexes en terme de temps de calcul, requièrent souvent une étape de précalcul hors-ligne. Les applications en temps réel, comme les jeux vidéo et la réalité virtuelle, se basent sur des techniques d’approximation et de précalcul pour atteindre des résultats réalistes. L’objectif de ce mémoire est l’investigation de différentes paramétrisations animées pour concevoir une technique d’approximation de rendu réaliste en temps réel. Notre investigation se concentre sur le rendu d’effets visuels appliqués à des personnages animés par modèle d’armature squelettique. Des paramétrisations combinant des données de mouvement et d’apparence nous permettent l’extraction de paramètres pour le processus en temps réel. Établir une dépendance linéaire entre le mouvement et l’apparence est ainsi au coeur de notre méthode. Nous nous concentrons sur l’occultation ambiante, où la simulation de l’occultation est causée par des objets à proximité bloquant la lumière environnante, jugée uniforme. L’occultation ambiante est une technique indépendante du point de vue, et elle est désormais essentielle pour le réalisme en temps réel. Nous examinons plusieurs paramétrisations qui traitent l’espace du maillage en fonction de l’information d’animation par squelette et/ou du maillage géométrique. Nous sommes capables d’approximer la réalité pour l’occultation ambiante avec une faible erreur. Notre technique pourrait également être étendue à d’autres effets visuels tels le rendu de la peau humaine (diffusion sous-surface), les changements de couleur dépendant du point de vue, les déformations musculaires, la fourrure ou encore les vêtements.While realistic rendering gains more popularity in industry, photorealistic and physically- based techniques often necessitate offline processing due to their computational complexity. Real-time applications, such as video games and virtual reality, rely mostly on approximation and precomputation techniques to achieve realistic results. The objective of this thesis is to investigate different animated parameterizations in order to devise a technique that can approximate realistic rendering results in real time. Our investigation focuses on rendering visual effects applied to skinned skeletonbased characters. Combined parameterizations of motion and appearance data are used to extract parameters that can be used in a real-time approximation. Trying to establish a linear dependency between motion and appearance is the basis of our method. We focus on ambient occlusion, a simulation of shadowing caused by objects that block ambient light. Ambient occlusion is a view-independent technique important for realism. We consider different parameterization techniques that treat the mesh space depending on skeletal animation information and/or mesh geometry. We are able to approximate ground-truth ambient occlusion with low error. Our technique can also be extended to different visual effects, such as rendering human skin (subsurface scattering), changes in color due to the view orientation, deformation of muscles, fur, or clothe

    Towards parameter-less 3D mesh segmentation

    Full text link
    This thesis focuses on the 3D mesh segmentation process. The research demonstrated how the process can be done in a parameterless approach which allows full automation with accurate results. Applications of this research include, but not limited to, 3D search engines, 3D character animation, robotics environment recognition, and augmented reality

    Tools for fluid simulation control in computer graphics

    Full text link
    L’animation basée sur la physique peut générer des systèmes aux comportements complexes et réalistes. Malheureusement, contrôler de tels systèmes est une tâche ardue. Dans le cas de la simulation de fluide, le processus de contrôle est particulièrement complexe. Bien que de nombreuses méthodes et outils ont été mis au point pour simuler et faire le rendu de fluides, trop peu de méthodes offrent un contrôle efficace et intuitif sur une simulation de fluide. Étant donné que le coût associé au contrôle vient souvent s’additionner au coût de la simulation, appliquer un contrôle sur une simulation à plus haute résolution rallonge chaque itération du processus de création. Afin d’accélérer ce processus, l’édition peut se faire sur une simulation basse résolution moins coûteuse. Nous pouvons donc considérer que la création d’un fluide contrôlé peut se diviser en deux phases: une phase de contrôle durant laquelle un artiste modifie le comportement d’une simulation basse résolution, et une phase d’augmentation de détail durant laquelle une version haute résolution de cette simulation est générée. Cette thèse présente deux projets, chacun contribuant à l’état de l’art relié à chacune de ces deux phases. Dans un premier temps, on introduit un nouveau système de contrôle de liquide représenté par un modèle particulaire. À l’aide de ce système, un artiste peut sélectionner dans une base de données une parcelle de liquide animé précalculée. Cette parcelle peut ensuite être placée dans une simulation afin d’en modifier son comportement. À chaque pas de simulation, notre système utilise la liste de parcelles actives afin de reproduire localement la vision de l’artiste. Une interface graphique intuitive a été développée, inspirée par les logiciels de montage vidéo, et permettant à un utilisateur non expert de simplement éditer une simulation de liquide. Dans un second temps, une méthode d’augmentation de détail est décrite. Nous proposons d’ajouter une étape supplémentaire de suivi après l’étape de projection du champ de vitesse d’une simulation de fumée eulérienne classique. Durant cette étape, un champ de perturbations de vitesse non-divergent est calculé, résultant en une meilleure correspondance des densités à haute et à basse résolution. L’animation de fumée résultante reproduit fidèlement l’aspect grossier de la simulation d’entrée, tout en étant augmentée à l’aide de détails simulés.Physics-based animation can generate dynamic systems of very complex and realistic behaviors. Unfortunately, controlling them is a daunting task. In particular, fluid simulation brings up particularly difficult problems to the control process. Although many methods and tools have been developed to convincingly simulate and render fluids, too few methods provide efficient and intuitive control over a simulation. Since control often comes with extra computations on top of the simulation cost, art-directing a high-resolution simulation leads to long iterations of the creative process. In order to shorten this process, editing could be performed on a faster, low-resolution model. Therefore, we can consider that the process of generating an art-directed fluid could be split into two stages: a control stage during which an artist modifies the behavior of a low-resolution simulation, and an upresolution stage during which a final high-resolution version of this simulation is driven. This thesis presents two projects, each one improving on the state of the art related to each of these two stages. First, we introduce a new particle-based liquid control system. Using this system, an artist selects patches of precomputed liquid animations from a database, and places them in a simulation to modify its behavior. At each simulation time step, our system uses these entities to control the simulation in order to reproduce the artist’s vision. An intuitive graphical user interface inspired by video editing tools has been developed, allowing a nontechnical user to simply edit a liquid animation. Second, a tracking solution for smoke upresolution is described. We propose to add an extra tracking step after the projection of a classical Eulerian smoke simulation. During this step, we solve for a divergence-free velocity perturbation field resulting in a better matching of the low-frequency density distribution between the low-resolution guide and the high-resolution simulation. The resulting smoke animation faithfully reproduces the coarse aspect of the low-resolution input, while being enhanced with simulated small-scale details
    • …
    corecore