15 research outputs found

    A Review of Dynamic Datasets for Facial Expression Research

    Get PDF
    Temporal dynamics have been increasingly recognized as an important component of facial expressions. With the need for appropriate stimuli in research and application, a range of databases of dynamic facial stimuli has been developed. The present article reviews the existing corpora and describes the key dimensions and properties of the available sets. This includes a discussion of conceptual features in terms of thematic issues in dataset construction as well as practical features which are of applied interest to stimulus usage. To identify the most influential sets, we further examine their citation rates and usage frequencies in existing studies. General limitations and implications for emotion research are noted and future directions for stimulus generation are outlined

    Dynamic Obstacle Clearing for Real-time Character Animation

    Get PDF
    This paper proposes a novel method to control virtual characters in dynamic environments. A virtual character is animated by a locomotion and jumping engine, enabling production of continuous parameterized motions. At any time during runtime, flat obstacles (e.g. a puddle of water) can be created and placed in front of a character. The method first decides whether the character is able to get around or jump over the obstacle. Then the motion parameters are accordingly modified. The transition from locomotion to jump is performed with an improved motion blending technique. While traditional blending approaches let the user choose the transition time and duration manually, our approach automatically controls transitions between motion patterns whose parameters are not known in advance. In addition, according to the animation context, blending operations are executed during a precise period of time to preserve specific physical properties. This ensures coherent movements over the parameter space of the original input motions. The initial locomotion type and speed are smoothly varied with respect to the required jump type and length. This variation is carefully computed in order to place the take-off foot as close to the created obstacle as possible

    A Motion Control Scheme for Animating Expressive Arm Movements

    Get PDF
    Current methods for figure animation involve a tradeoff between the level of realism captured in the movements and the ease of generating the animations. We introduce a motion control paradigm that circumvents this tradeoff-it provides the ability to generate a wide range of natural-looking movements with minimal user labor. Effort, which is one part of Rudolf Laban\u27s system for observing and analyzing movement, describes the qualitative aspects of movement. Our motion control paradigm simplifies the generation of expressive movements by proceduralizing these qualitative aspects to hide the non-intuitive, quantitative aspects of movement. We build a model of Effort using a set of kinematic movement parameters that defines how a figure moves between goal keypoints. Our motion control scheme provides control through Effort\u27s four dimensional system of textual descriptors, providing a level of control thus far missing from behavioral animation systems and offering novel specification and editing capabilities on top of traditional keyframing and inverse kinematics methods. Since our Effort model is inexpensive computationally, Effort-based motion control systems can work in real-time. We demonstrate our motion control scheme by implementing EMOTE (Expressive MOTion Engine), a character animation module for expressive arm movements. EMOTE works with inverse kinematics to control the qualitative aspects of end-effector specified movements. The user specifies general movements by entering a sequence of goal positions for each hand. The user then expresses the essence of the movement by adjusting sliders for the Effort motion factors: Space, Weight, Time, and Flow. EMOTE produces a wide range of expressive movements, provides an easy-to-use interface (that is more intuitive than joint angle interpolation curves or physical parameters), features interactive editing, and real-time motion generation

    3D face modelling from sparse data

    Get PDF
    EThOS - Electronic Theses Online ServiceGBUnited Kingdo

    Motion Planning : from Digital Actors to Humanoid Robots

    Get PDF
    Le but de ce travail est de développer des algorithmes de planification de mouvement pour des figures anthropomorphes en tenant compte de la géométrie, de la cinématique et de la dynamique du mécanisme et de son environnement. Par planification de mouvement, on entend la capacité de donner des directives à un niveau élevé et de les transformer en instructions de bas niveau qui produiront une séquence de valeurs articulaires qui reproduissent les mouvements humains. Ces instructions doivent considérer l'évitement des obstacles dans un environnement qui peut être plus au moins contraint. Ceci a comme consequence que l'on peut exprimer des directives comme “porte ce plat de la table jusqu'ac'estu coin du piano”, qui seront ensuite traduites en une série de buts intermédiaires et de contraintes qui produiront les mouvements appropriés des articulations du robot, de façon a effectuer l'action demandée tout en evitant les obstacles dans la chambre. Nos algorithmes se basent sur l'observation que les humains ne planifient pas des mouvements précis pour aller à un endroit donné. On planifie grossièrement la direction de marche et, tout en avançant, on exécute les mouvements nécessaires des articulations afin de nous mener à l'endroit voulu. Nous avons donc cherché à concevoir des algorithmes au sein d'un tel paradigme, algorithmes qui: 1. Produisent un chemin sans collision avec une version réduite du mécanisme et qui le mènent au but spécifié. 2. Utilisent les contrôleurs disponibles pour générer un mouvement qui assigne des valeurs à chacune des articulations du mécanisme pour suivre le chemin trouvé précédemment. 3. Modifient itérativement ces trajectoires jusqu'à ce que toutes les contraintes géométriques, cinématiques et dynamiques soient satisfaites. Dans ce travail nous appliquons cette approche à trois étages au problème de la planification de mouvements pour des figures anthropomorphes qui manipulent des objets encombrants tout en marchant. Dans le processus, plusieurs problèmes intéressants, ainsi que des propositions pour les résoudre, sont présentés. Ces problèmes sont principalement l'évitement tri-dimensionnel des obstacles, la manipulation des objets à deux mains, la manipulation coopérative des objets et la combinaison de comportements hétérogènes. La contribution principale de ce travail est la modélisation du problème de la génération automatique des mouvements de manipulation et de locomotion. Ce modèle considère les difficultés exprimées ci dessus, dans les contexte de mécanismes bipèdes. Trois principes fondent notre modèle: une décomposition fonctionnelle des membres du mécanisme, un modèle de manipulation coopérative et, un modéle simplifié des facultés de déplacement du mécanisme dans son environnement.Ce travail est principalement et surtout, un travail de synthèse. Nous nous servons des techniques disponibles pour commander la locomotion des mécanismes bipèdes (contrôleurs) provenant soit de l'animation par ordinateur, soit de la robotique humanoïde, et nous les relions dans un planificateur des mouvements original. Ce planificateur de mouvements est agnostique vis-à-vis du contrôleur utilisé, c'est-à-dire qu'il est capable de produire des mouvements libres de collision avec n'importe quel contrôleur tandis que les entrées et sorties restent compatibles. Naturellement, l'exécution de notre planificateur dépend en grand partie de la qualité du contrôleur utilisé. Dans cette thèse, le planificateur de mouvement est relié à différents contrôleurs et ses bonnes performances sont validées avec des mécanismes différents, tant virtuels que physiques. Ce travail à été fait dans le cadre des projets de recherche communs entre la France, la Russie et le Japon, où nous avons fourni le cadre de planification de mouvement à ses différents contrôleurs. Plusieurs publications issues de ces collaborations ont été présentées dans des conférences internationales. Ces résultats sont compilés et présentés dans cette thèse, et le choix des techniques ainsi que les avantages et inconvénients de notre approche sont discutés. ABSTRACT : The goal of this work is to develop motion planning algorithms for human-like figures taking into account the geometry, kinematics and dynamics of the mechanism and its environment. By motion planning it is understood the ability to specify high-level directives and transform them into low-level instructions for the articulations of the human-like figure. This is usually done while considering obstacle avoidance within the environment. This results in one being able to express directives as “carry this plate from the table to the piano corner” and have them translate into a series of goals and constraints that result in the pertinent motions from the robot's articulations in such a way as to carry out the action while avoiding collisions with the obstacles in the room. Our algorithms are based on the observation that humans do not plan their exact motions when getting to a location. We roughly plan our direction and, as we advance, we execute the motions needed to get to the desired place. This has led us to design algorithms that: 1. Produce a rough collision free path that takes a simplified model of the mechanism to the desired location. 2. Use available controllers to generate a trajectory that assigns values to each of the mechanism's articulations to follow the path. 3. Modify iteratively these trajectories until all the geometric, kinematic and dynamic constraints of the problem are satisfied.Throughout this work, we apply this three-stage approach with the problem of generating motions for human-like figures that manipulate bulky objects while walking. In the process, several interesting problems and their solution are brought into focus. These problems are, three- imensional collision avoidance, two-hand object manipulation, cooperative manipulation among several characters or robots and the combination of different behaviors. The main contribution of this work is the modeling of the automatic generation of cooperative manipulation motions. This model considers the above difficulties, all in the context of bipedal walking mechanisms. Three principles inform the model: a functional decomposition of the mechanism's limbs, a model for cooperative manipulation and, a simplified model to represent the mechanism when generating the rough path. This work is mainly and above all, one of synthesis. We make use of available techniques for controlling locomotion of bipedal mechanisms (controllers), from the fields of computer graphics and robotics, and connect them to a novel motion planner. This motion planner is controller-agnostic, that is, it is able to produce collision-free motions with any controller, despite whatever errors introduced by the controller itself. Of course, the performance of our motion planner depends on the quality of the used controller. In this thesis, the motion planner, connected to different controllers, is used and tested in different mechanisms, both virtual and physical. This in the context of different research projects in France, Russia and Japan, where we have provided the motion planning framework to their controllers. Several papers in peer-reviewed international conferences have resulted from these collaborations. The present work compiles these results and provides a more comprehensive and detailed depiction of the system and its benefits, both when applied to different mechanisms and compared to alternative approache

    Imitation and social learning for synthetic characters

    Get PDF
    Thesis (S.M.)--Massachusetts Institute of Technology, School of Architecture and Planning, Program in Media Arts and Sciences, 2004.Includes bibliographical references (p. 137-149).We want to build animated characters and robots capable of rich social interactions with humans and each other, and who are able to learn by observing those around them. An increasing amount of evidence suggests that, in human infants, the ability to learn by watching others, and in particular, the ability to imitate, could be crucial precursors to the development of appropriate social behavior, and ultimately the ability to reason about the thoughts, intents, beliefs, and desires of others. We have created a number of imitative characters and robots, the latest of which is Max T. Mouse, an anthropomorphic animated mouse character who is able to observe the actions he sees his friend Morris Mouse performing, and compare them to the actions he knows how to perform himself. This matching process allows Max to accurately imitate Morris's gestures and actions, even when provided with limited synthetic visual input. Furthermore, by using his own perception, motor, and action systems as models for the behavioral and perceptual capabilities of others (a process known as Simulation Theory in the cognitive literature), Max can begin to identify simple goals and motivations for Morris's behavior, an important step towards developing characters with a full theory of mind. Finally, Max can learn about unfamiliar objects in his environment, such as food and toys, by observing and correctly interpreting Morris's interactions with these objects, demonstrating his ability to take advantage of socially acquired information.by Daphna Buchsbaum.S.M

    Automating the conversion of natural language fiction to multi-modal 3D animated virtual environments

    Get PDF
    Popular fiction books describe rich visual environments that contain characters, objects, and behaviour. This research develops automated processes for converting text sourced from fiction books into animated virtual environments and multi-modal films. This involves the analysis of unrestricted natural language fiction to identify appropriate visual descriptions, and the interpretation of the identified descriptions for constructing animated 3D virtual environments. The goal of the text analysis stage is the creation of annotated fiction text, which identifies visual descriptions in a structured manner. A hierarchical rule-based learning system is created that induces patterns from example annotations provided by a human, and uses these for the creation of additional annotations. Patterns are expressed as tree structures that abstract the input text on different levels according to structural (token, sentence) and syntactic (parts-of-speech, syntactic function) categories. Patterns are generalized using pair-wise merging, where dissimilar sub-trees are replaced with wild-cards. The result is a small set of generalized patterns that are able to create correct annotations. A set of generalized patterns represents a model of an annotator's mental process regarding a particular annotation category. Annotated text is interpreted automatically for constructing detailed scene descriptions. This includes identifying which scenes to visualize, and identifying the contents and behaviour in each scene. Entity behaviour in a 3D virtual environment is formulated using time-based constraints that are automatically derived from annotations. Constraints are expressed as non-linear symbolic functions that restrict the trajectories of a pair of entities over a continuous interval of time. Solutions to these constraints specify precise behaviour. We create an innovative quantified constraint optimizer for locating sound solutions, which uses interval arithmetic for treating time and space as contiguous quantities. This optimization method uses a technique of constraint relaxation and tightening that allows solution approximations to be located where constraint systems are inconsistent (an ability not previously explored in interval-based quantified constraint solving). 3D virtual environments are populated by automatically selecting geometric models or procedural geometry-creation methods from a library. 3D models are animated according to trajectories derived from constraint solutions. The final animated film is sequenced using a range of modalities including animated 3D graphics, textual subtitles, audio narrations, and foleys. Hierarchical rule-based learning is evaluated over a range of annotation categories. Models are induced for different categories of annotation without modifying the core learning algorithms, and these models are shown to be applicable to different types of books. Models are induced automatically with accuracies ranging between 51.4% and 90.4%, depending on the category. We show that models are refined if further examples are provided, and this supports a boot-strapping process for training the learning mechanism. The task of interpreting annotated fiction text and populating 3D virtual environments is successfully automated using our described techniques. Detailed scene descriptions are created accurately, where between 83% and 96% of the automatically generated descriptions require no manual modification (depending on the type of description). The interval-based quantified constraint optimizer fully automates the behaviour specification process. Sample animated multi-modal 3D films are created using extracts from fiction books that are unrestricted in terms of complexity or subject matter (unlike existing text-to-graphics systems). These examples demonstrate that: behaviour is visualized that corresponds to the descriptions in the original text; appropriate geometry is selected (or created) for visualizing entities in each scene; sequences of scenes are created for a film-like presentation of the story; and that multiple modalities are combined to create a coherent multi-modal representation of the fiction text. This research demonstrates that visual descriptions in fiction text can be automatically identified, and that these descriptions can be converted into corresponding animated virtual environments. Unlike existing text-to-graphics systems, we describe techniques that function over unrestricted natural language text and perform the conversion process without the need for manually constructed repositories of world knowledge. This enables the rapid production of animated 3D virtual environments, allowing the human designer to focus on creative aspects

    Generating whole body movements for dynamics anthropomorphic systems under constraints

    Get PDF
    Cette thèse étudie la question de la génération de mouvements corps-complet pour des systèmes anthropomorphes. Elle considère le problème de la modélisation et de la commande en abordant la question difficile de la génération de mouvements ressemblant à ceux de l'homme. En premier lieu, un modèle dynamique du robot humanoïde HRP-2 est élaboré à partir de l'algorithme récursif de Newton-Euler pour les vecteurs spatiaux. Un nouveau schéma de commande dynamique est ensuite développé, en utilisant une cascade de programmes quadratiques (QP) optimisant des fonctions coûts et calculant les couples de commande en satisfaisant des contraintes d'égalité et d'inégalité. La cascade de problèmes quadratiques est définie par une pile de tâches associée à un ordre de priorité. Nous proposons ensuite une formulation unifiée des contraintes de contacts planaires et nous montrons que la méthode proposée permet de prendre en compte plusieurs contacts non coplanaires et généralise la contrainte usuelle du ZMP dans le cas où seulement les pieds sont en contact avec le sol. Nous relions ensuite les algorithmes de génération de mouvement issus de la robotique aux outils de capture du mouvement humain en développant une méthode originale de génération de mouvement visant à imiter le mouvement humain. Cette méthode est basée sur le recalage des données capturées et l'édition du mouvement en utilisant le solveur hiérarchique précédemment introduit et la définition de tâches et de contraintes dynamiques. Cette méthode originale permet d'ajuster un mouvement humain capturé pour le reproduire fidèlement sur un humanoïde en respectant sa propre dynamique. Enfin, dans le but de simuler des mouvements qui ressemblent à ceux de l'homme, nous développons un modèle anthropomorphe ayant un nombre de degrés de liberté supérieur à celui du robot humanoïde HRP2. Le solveur générique est utilisé pour simuler le mouvement sur ce nouveau modèle. Une série de tâches est définie pour décrire un scénario joué par un humain. Nous montrons, par une simple analyse qualitative du mouvement, que la prise en compte du modèle dynamique permet d'accroitre naturellement le réalisme du mouvement.This thesis studies the question of whole body motion generation for anthropomorphic systems. Within this work, the problem of modeling and control is considered by addressing the difficult issue of generating human-like motion. First, a dynamic model of the humanoid robot HRP-2 is elaborated based on the recursive Newton-Euler algorithm for spatial vectors. A new dynamic control scheme is then developed adopting a cascade of quadratic programs (QP) optimizing the cost functions and computing the torque control while satisfying equality and inequality constraints. The cascade of the quadratic programs is defined by a stack of tasks associated to a priority order. Next, we propose a unified formulation of the planar contact constraints, and we demonstrate that the proposed method allows taking into account multiple non coplanar contacts and generalizes the common ZMP constraint when only the feet are in contact with the ground. Then, we link the algorithms of motion generation resulting from robotics to the human motion capture tools by developing an original method of motion generation aiming at the imitation of the human motion. This method is based on the reshaping of the captured data and the motion editing by using the hierarchical solver previously introduced and the definition of dynamic tasks and constraints. This original method allows adjusting a captured human motion in order to reliably reproduce it on a humanoid while respecting its own dynamics. Finally, in order to simulate movements resembling to those of humans, we develop an anthropomorphic model with higher number of degrees of freedom than the one of HRP-2. The generic solver is used to simulate motion on this new model. A sequence of tasks is defined to describe a scenario played by a human. By a simple qualitative analysis of motion, we demonstrate that taking into account the dynamics provides a natural way to generate human-like movements

    Quantification of knee extensor muscle forces: a multimodality approach

    Get PDF
    Given the growing interest of using musculoskeletal (MSK) models in a large number of clinical applications for quantifying the internal loading of the human MSK system, verification and validation of the model’s predictions, especially at the knee joint, have remained as one of the biggest challenges in the use of the models as clinical tools. This thesis proposes a methodology for more accurate quantification of knee extensor forces by exploring different experimental and modelling techniques that can be used to enhance the process of verification and validation of the knee joint model within the MSK models for transforming the models to a viable clinical tool. In this methodology, an experimental protocol was developed for simultaneous measurement of the knee joint motion, torques, external forces and muscular activation during an isolated knee extension exercise. This experimental protocol was tested on a cohort of 11 male subjects and the measurements were used to quantify knee extensor forces using two different MSK models representing a simplified model of the knee extensor mechanism and a previously-developed three-dimensional MSK model of the lower limb. The quantified knee extensor forces from the MSK models were then compared to evaluate the performance of the models for quantifying knee extensor forces. The MSK models were also used to investigate the sensitivity of the calculated knee extensor forces to key modelling parameters of the knee including the method of quantifying the knee centre of rotation and the effect of joint translation during motion. In addition, the feasibility of an emerging ultrasound-based imaging technique (shear wave elastography) for direct quantification of the physiologically-relevant musculotendon forces was investigated. The results in this thesis showed that a simplified model of the knee can be reliably used during a controlled planar activity as a computationally-fast and effective tool for hierarchical verification of the knee joint model in optimisation-based large-scale MSK models to provide more confidence in the outputs of the models. Furthermore, the calculation of knee extensor muscle forces has been found to be sensitive to knee joint translation (moving centre of rotation of the knee), highlighting the importance of this modelling parameter for quantifying physiologically-realistic knee muscle forces in the MSK models. It was also demonstrated how the movement of the knee axis of rotation during motion can be used as an intuitive tool for understanding the functional anatomy of the knee joint. Moreover, the findings in this thesis indicated that the shear wave elastography technique can be potentially used as a novel method for direct quantification of the physiologically-relevant musculotendon forces for independent validation of the predictions of musculotendon forces from the MSK models.Open Acces
    corecore