15,287 research outputs found

    Spectral Style Transfer for Human Motion between Independent Actions

    Get PDF
    Human motion is complex and difficult to synthesize realistically. Automatic style transfer to transform the mood or identity of a character's motion is a key technology for increasing the value of already synthesized or captured motion data. Typically, state-of-the-art methods require all independent actions observed in the input to be present in a given style database to perform realistic style transfer. We introduce a spectral style transfer method for human motion between independent actions, thereby greatly reducing the required effort and cost of creating such databases. We leverage a spectral domain representation of the human motion to formulate a spatial correspondence free approach. We extract spectral intensity representations of reference and source styles for an arbitrary action, and transfer their difference to a novel motion which may contain previously unseen actions. Building on this core method, we introduce a temporally sliding window filter to perform the same analysis locally in time for heterogeneous motion processing. This immediately allows our approach to serve as a style database enhancement technique to fill-in non-existent actions in order to increase previous style transfer method's performance. We evaluate our method both via quantitative experiments, and through administering controlled user studies with respect to previous work, where significant improvement is observed with our approach

    Synthesizing Human Actions with Emotion

    Get PDF
    Realistic synthesis of human actions is a challenging problem. This thesis investigates the problem of synthesizing actions, with individual variability, under different emotions. Current actions/gesture synthesis, understanding and recognition models do not provide a general framework for synthesizing an extensive range of actions over a large range of emotions.The literature on spectral style transfer provides a plethora of viable approaches for transferring the style of action learned from one individual to another. Our idea is to consider an emotion as a style then use a style transfer algorithm for transferring an emotion from one action to another. This allows us to synthesize any action over a large range of emotions. Experiments reported in this thesis are based on genarating18 actions with five emotions using the Kinect skeleton. The quality of the synthesized actions over time is evaluated through a subjective perception test, which is a standard in the domain of gesture synthesis

    Using a Cognitive Architecture for Opponent Target Prediction

    No full text
    One of the most important aspects of a compelling game AI is that it anticipates the player’s actions and responds to them in a convincing manner. The first step towards doing this is to understand what the player is doing and predict their possible future actions. In this paper we show an approach where the AI system focusses on testing hypotheses made about the player’s actions using an implementation of a cognitive architecture inspired by the simulation theory of mind. The application used in this paper is to predict the target that the player is heading towards, in an RTS-style game. We improve the prediction accuracy and reduce the number of hypotheses needed by using path planning and path clustering

    Deformable Shape Completion with Graph Convolutional Autoencoders

    Full text link
    The availability of affordable and portable depth sensors has made scanning objects and people simpler than ever. However, dealing with occlusions and missing parts is still a significant challenge. The problem of reconstructing a (possibly non-rigidly moving) 3D object from a single or multiple partial scans has received increasing attention in recent years. In this work, we propose a novel learning-based method for the completion of partial shapes. Unlike the majority of existing approaches, our method focuses on objects that can undergo non-rigid deformations. The core of our method is a variational autoencoder with graph convolutional operations that learns a latent space for complete realistic shapes. At inference, we optimize to find the representation in this latent space that best fits the generated shape to the known partial input. The completed shape exhibits a realistic appearance on the unknown part. We show promising results towards the completion of synthetic and real scans of human body and face meshes exhibiting different styles of articulation and partiality.Comment: CVPR 201

    A note on brain actuated spelling with the Berlin brain-computer interface

    Get PDF
    Brain-Computer Interfaces (BCIs) are systems capable of decoding neural activity in real time, thereby allowing a computer application to be directly controlled by the brain. Since the characteristics of such direct brain-tocomputer interaction are limited in several aspects, one major challenge in BCI research is intelligent front-end design. Here we present the mental text entry application ‘Hex-o-Spell’ which incorporates principles of Human-Computer Interaction research into BCI feedback design. The system utilises the high visual display bandwidth to help compensate for the extremely limited control bandwidth which operates with only two mental states, where the timing of the state changes encodes most of the information. The display is visually appealing, and control is robust. The effectiveness and robustness of the interface was demonstrated at the CeBIT 2006 (world’s largest IT fair) where two subjects operated the mental text entry system at a speed of up to 7.6 char/min

    Combining computer game-based behavioural experiments with high-density EEG and infrared gaze tracking

    Get PDF
    Rigorous, quantitative examination of therapeutic techniques anecdotally reported to have been successful in people with autism who lack communicative speech will help guide basic science toward a more complete characterisation of the cognitive profile in this underserved subpopulation, and show the extent to which theories and results developed with the high-functioning subpopulation may apply. This study examines a novel therapy, the "Rapid Prompting Method" (RPM). RPM is a parent-developed communicative and educational therapy for persons with autism who do not speak or who have difficulty using speech communicatively.The technique aims to develop a means of interactive learning by pointing amongst multiple-choice options presented at different locations in space, with the aid of sensory "prompts" which evoke a response without cueing any specific response option. The prompts are meant to draw and to maintain attention to the communicative task–making the communicative and educational content coincident with the most physically salient, attention-capturing stimulus – and to extinguish the sensory–motor preoccupations with which the prompts compete.ideo-recorded RPM sessions with nine autistic children ages 8–14years who lacked functional communicative speech were coded for behaviours of interest

    Synthesizing Skeletal Motion and Physiological Signals as a Function of a Virtual Human's Actions and Emotions

    Full text link
    Round-the-clock monitoring of human behavior and emotions is required in many healthcare applications which is very expensive but can be automated using machine learning (ML) and sensor technologies. Unfortunately, the lack of infrastructure for collection and sharing of such data is a bottleneck for ML research applied to healthcare. Our goal is to circumvent this bottleneck by simulating a human body in virtual environment. This will allow generation of potentially infinite amounts of shareable data from an individual as a function of his actions, interactions and emotions in a care facility or at home, with no risk of confidentiality breach or privacy invasion. In this paper, we develop for the first time a system consisting of computational models for synchronously synthesizing skeletal motion, electrocardiogram, blood pressure, respiration, and skin conductance signals as a function of an open-ended set of actions and emotions. Our experimental evaluations, involving user studies, benchmark datasets and comparison to findings in the literature, show that our models can generate skeletal motion and physiological signals with high fidelity. The proposed framework is modular and allows the flexibility to experiment with different models. In addition to facilitating ML research for round-the-clock monitoring at a reduced cost, the proposed framework will allow reusability of code and data, and may be used as a training tool for ML practitioners and healthcare professionals
    corecore