14 research outputs found

    Extracting Motion Primitives from Natural Handwriting Data

    Get PDF
    Institute for Adaptive and Neural ComputationHumans and animals can plan and execute movements much more adaptably and reliably than current computers can calculate robotic limb trajectories. Over recent decades, it has been suggested that our brains use motor primitives as blocks to build up movements. In broad terms a primitive is a segment of pre-optimised movement allowing a simplified movement planning solution. This thesis explores a generative model of handwriting based upon the concept of motor primitives. Unlike most primitive extraction studies, the primitives here are time extended blocks that are superimposed with character specific offsets to create a pen trajectory. This thesis shows how handwriting can be represented using a simple fixed function superposition model, where the variation in the handwriting arises from timing variation in the onset of the functions. Furthermore, it is shown how handwriting style variations could be due to primitive function differences between individuals, and how the timing code could provide a style invariant representation of the handwriting. The spike timing representation of the pen movements provides an extremely compact code, which could resemble internal spiking neural representations in the brain. The model proposes an novel way to infer primitives in data, and the proposed formalised probabilistic model allows informative priors to be introduced providing a more accurate inference of primitive shape and timing

    Understanding of Object Manipulation Actions Using Human Multi-Modal Sensory Data

    Full text link
    Object manipulation actions represent an important share of the Activities of Daily Living (ADLs). In this work, we study how to enable service robots to use human multi-modal data to understand object manipulation actions, and how they can recognize such actions when humans perform them during human-robot collaboration tasks. The multi-modal data in this study consists of videos, hand motion data, applied forces as represented by the pressure patterns on the hand, and measurements of the bending of the fingers, collected as human subjects performed manipulation actions. We investigate two different approaches. In the first one, we show that multi-modal signal (motion, finger bending and hand pressure) generated by the action can be decomposed into a set of primitives that can be seen as its building blocks. These primitives are used to define 24 multi-modal primitive features. The primitive features can in turn be used as an abstract representation of the multi-modal signal and employed for action recognition. In the latter approach, the visual features are extracted from the data using a pre-trained image classification deep convolutional neural network. The visual features are subsequently used to train the classifier. We also investigate whether adding data from other modalities produces a statistically significant improvement in the classifier performance. We show that both approaches produce a comparable performance. This implies that image-based methods can successfully recognize human actions during human-robot collaboration. On the other hand, in order to provide training data for the robot so it can learn how to perform object manipulation actions, multi-modal data provides a better alternative

    Signature features with the visibility transformation

    Get PDF
    The signature in rough path theory provides a graduated summary of a path through an examination of the effects of its increments. Inspired by recent developments of signature features in the context of machine learning, we explore a transformation that is able to embed the effect of the absolute position of the data stream into signature features. This unified feature is particularly effective for its simplifying role in allowing the signature feature set to accommodate nonlinear functions of absolute and relative values

    Extracting motion primitives from natural handwriting data

    No full text
    Abstract. For the past 10 years it has become clear that biological movement is made up of sub-routine type blocks, or motor primitives, with a central controller timing the activation of these blocks, creating synergies of muscle activation. This paper shows that it is possible to use a factorial hidden Markov model to infer primitives in handwriting data. These primitives are not predefined in terms of location of occurrence within the handwriting, and they are not limited or defined by a particular character set. Also, the variation in the data can to a large extent be explained by timing variation in the triggering of the primitives. Once an appropriate set of primitives has been inferred, the characters can be represented as a set of timings of primitive activations, along with variances, giving a very compact representation of the character. Separating the motor system into a motor primitive part, and a timing control gives us a possible insight into how we might create scribbles on paper.

    Extracting motion primitives from natural handwriting data

    No full text
    Humans and animals can plan and execute movements much more adaptably and reliably than current computers can calculate robotic limb trajectories. Over recent decades, it has been suggested that our brains use motor primitives as blocks to build up movements. In broad terms a primitive is a segment of pre-optimised movement allowing a simplified movement planning solution. This thesis explores a generative model of handwriting based upon the concept of motor primitives. Unlike most primitive extraction studies, the primitives here are time extended blocks that are superimposed with character specific offsets to create a pen trajectory. This thesis shows how handwriting can be represented using a simple fixed function superposition model, where the variation in the handwriting arises from timing variation in the onset of the functions. Furthermore, it is shown how handwriting style variations could be due to primitive function differences between individuals, and how the timing code could provide a style invariant representation of the handwriting. The spike timing representation of the pen movements provides an extremely compact code, which could resemble internal spiking neural representations in the brain. The model proposes an novel way to infer primitives in data, and the proposed formalised probabilistic model allows informative priors to be introduced providing a more accurate inference of primitive shape and timing.EThOS - Electronic Theses Online ServiceGBUnited Kingdo

    Intention Inference and Decision Making with Hierarchical Gaussian Process Dynamics Models

    Get PDF
    Anticipation is crucial for fluent human-robot interaction, which allows a robot to independently coordinate its actions with human beings in joint activities. An anticipatory robot relies on a predictive model of its human partners, and selects its own action according to the model's predictions. Intention inference and decision making are key elements towards such anticipatory robots. In this thesis, we present a machine-learning approach to intention inference and decision making, based on Hierarchical Gaussian Process Dynamics Models (H-GPDMs). We first introduce the H-GPDM, a class of generic latent-variable dynamics models. The H-GPDM represents the generative process of complex human movements that are directed by exogenous driving factors. Incorporating the exogenous variables in the dynamics model, the H-GPDM achieves improved interpretation, analysis, and prediction of human movements. While exact inference of the exogenous variables and the latent states is intractable, we introduce an approximate method using variational Bayesian inference, and demonstrate the merits of the H-GPDM in three different applications of human movement analysis. The H-GPDM lays a foundation for the following studies on intention inference and decision making. Intention inference is an essential step towards anticipatory robots. For this purpose, we consider a special case of the H-GPDM, the Intention-Driven Dynamics Model (IDDM), which considers the human partners' intention as exogenous driving factors. The IDDM is applicable to intention inference from observed movements using Bayes' theorem, where the latent state variables are marginalized out. As most robotics applications are subject to real-time constraints, we introduce an efficient online algorithm that allows for real-time intention inference. We show that the IDDM achieved state-of-the-art performance in intention inference using two human-robot interaction scenarios, i.e., target prediction for robot table tennis and action recognition for interactive robots. Decision making based on a time series of predictions allows a robot to be proactive in its action selection, which involves a trade-off between the accuracy and confidence of the prediction and the time for executing a selected action. To address the problem of action selection and optimal timing for initiating the movement, we formulate the anticipatory action selection using Partially Observable Markov Decision Process, where the H-GPDM is adopted to update belief state and to estimate transition model. We present two approaches to policy learning and decision making, and show their effectiveness using human-robot table tennis. In addition, we consider decision making solely based on the preference of the human partners, where observations are not sufficient for reliable intention inference. We formulate it as a repeated game and present a learning approach to safe strategies that exploit the humans' preferences. The learned strategy enables action selection when reliable intention inference is not available due to insufficient observation, e.g., for a robot to return served balls from a human table tennis player. In this thesis, we use human-robot table tennis as a running example, where a key bottleneck is the limited amount of time for executing a hitting movement. Movement initiation usually requires an early decision on the type of action, such as a forehand or backhand hitting movement, at least 80ms before the opponent has hit the ball. The robot, therefore, needs to be anticipatory and proactive of the opponent's intended target. Using the proposed methods, the robot can predict the intended target of the opponent and initiate an appropriate hitting movement according to the prediction. Experimental results show that the proposed intention inference and decision making methods can substantially enhance the capability of the robot table tennis player, using both a physically realistic simulation and a real Barrett WAM robot arm with seven degrees of freedom
    corecore