With the advancement of robotics, machine learning, and machine perception,
increasingly more robots will enter human environments to assist with daily
tasks. However, dynamically-changing human environments requires reactive
motion plans. Reactivity can be accomplished through replanning, e.g.
model-predictive control, or through a reactive feedback policy that modifies
on-going behavior in response to sensory events. In this paper, we investigate
how to use machine learning to add reactivity to a previously learned nominal
skilled behavior. We approach this by learning a reactive modification term for
movement plans represented by nonlinear differential equations. In particular,
we use dynamic movement primitives (DMPs) to represent a skill and a neural
network to learn a reactive policy from human demonstrations. We use the well
explored domain of obstacle avoidance for robot manipulation as a test bed. Our
approach demonstrates how a neural network can be combined with physical
insights to ensure robust behavior across different obstacle settings and
movement durations. Evaluations on an anthropomorphic robotic system
demonstrate the effectiveness of our work.Comment: 8 pages, accepted to be published at ICRA 2017 conferenc