184 research outputs found
Neural Dynamic Movement Primitives -- a survey
One of the most important challenges in robotics is producing accurate
trajectories and controlling their dynamic parameters so that the robots can
perform different tasks. The ability to provide such motion control is closely
related to how such movements are encoded. Advances on deep learning have had a
strong repercussion in the development of novel approaches for Dynamic Movement
Primitives. In this work, we survey scientific literature related to Neural
Dynamic Movement Primitives, to complement existing surveys on Dynamic Movement
Primitives
Deep Predictive Policy Training using Reinforcement Learning
Skilled robot task learning is best implemented by predictive action policies
due to the inherent latency of sensorimotor processes. However, training such
predictive policies is challenging as it involves finding a trajectory of motor
activations for the full duration of the action. We propose a data-efficient
deep predictive policy training (DPPT) framework with a deep neural network
policy architecture which maps an image observation to a sequence of motor
activations. The architecture consists of three sub-networks referred to as the
perception, policy and behavior super-layers. The perception and behavior
super-layers force an abstraction of visual and motor data trained with
synthetic and simulated training samples, respectively. The policy super-layer
is a small sub-network with fewer parameters that maps data in-between the
abstracted manifolds. It is trained for each task using methods for policy
search reinforcement learning. We demonstrate the suitability of the proposed
architecture and learning framework by training predictive policies for skilled
object grasping and ball throwing on a PR2 robot. The effectiveness of the
method is illustrated by the fact that these tasks are trained using only about
180 real robot attempts with qualitative terminal rewards.Comment: This work is submitted to IEEE/RSJ International Conference on
Intelligent Robots and Systems 2017 (IROS2017
MILD: Multimodal Interactive Latent Dynamics for Learning Human-Robot Interaction
Modeling interaction dynamics to generate robot trajectories that enable a
robot to adapt and react to a human's actions and intentions is critical for
efficient and effective collaborative Human-Robot Interactions (HRI). Learning
from Demonstration (LfD) methods from Human-Human Interactions (HHI) have shown
promising results, especially when coupled with representation learning
techniques. However, such methods for learning HRI either do not scale well to
high dimensional data or cannot accurately adapt to changing via-poses of the
interacting partner. We propose Multimodal Interactive Latent Dynamics (MILD),
a method that couples deep representation learning and probabilistic machine
learning to address the problem of two-party physical HRIs. We learn the
interaction dynamics from demonstrations, using Hidden Semi-Markov Models
(HSMMs) to model the joint distribution of the interacting agents in the latent
space of a Variational Autoencoder (VAE). Our experimental evaluations for
learning HRI from HHI demonstrations show that MILD effectively captures the
multimodality in the latent representations of HRI tasks, allowing us to decode
the varying dynamics occurring in such tasks. Compared to related work, MILD
generates more accurate trajectories for the controlled agent (robot) when
conditioned on the observed agent's (human) trajectory. Notably, MILD can learn
directly from camera-based pose estimations to generate trajectories, which we
then map to a humanoid robot without the need for any additional training.Comment: Accepted at the IEEE-RAS International Conference on Humanoid Robots
(Humanoids) 202
Learning from Demonstration with Weakly Supervised Disentanglement
Robotic manipulation tasks, such as wiping with a soft sponge, require
control from multiple rich sensory modalities. Human-robot interaction, aimed
at teaching robots, is difficult in this setting as there is potential for
mismatch between human and machine comprehension of the rich data streams. We
treat the task of interpretable learning from demonstration as an optimisation
problem over a probabilistic generative model. To account for the
high-dimensionality of the data, a high-capacity neural network is chosen to
represent the model. The latent variables in this model are explicitly aligned
with high-level notions and concepts that are manifested in a set of
demonstrations. We show that such alignment is best achieved through the use of
labels from the end user, in an appropriately restricted vocabulary, in
contrast to the conventional approach of the designer picking a prior over the
latent variables. Our approach is evaluated in the context of two table-top
robot manipulation tasks performed by a PR2 robot -- that of dabbing liquids
with a sponge (forcefully pressing a sponge and moving it along a surface) and
pouring between different containers. The robot provides visual information,
arm joint positions and arm joint efforts. We have made videos of the tasks and
data available - see supplementary materials at:
https://sites.google.com/view/weak-label-lfd.Comment: 18 pages, 16 figures, accepted at the International Conference on
Learning Representations (ICLR) 2021, supplementary website at
https://sites.google.com/view/weak-label-lf
- …