838 research outputs found
Learning-based Feedback Controller for Deformable Object Manipulation
In this paper, we present a general learning-based framework to automatically
visual-servo control the position and shape of a deformable object with unknown
deformation parameters. The servo-control is accomplished by learning a
feedback controller that determines the robotic end-effector's movement
according to the deformable object's current status. This status encodes the
object's deformation behavior by using a set of observed visual features, which
are either manually designed or automatically extracted from the robot's sensor
stream. A feedback control policy is then optimized to push the object toward a
desired featured status efficiently. The feedback policy can be learned either
online or offline. Our online policy learning is based on the Gaussian Process
Regression (GPR), which can achieve fast and accurate manipulation and is
robust to small perturbations. An offline imitation learning framework is also
proposed to achieve a control policy that is robust to large perturbations in
the human-robot interaction. We validate the performance of our controller on a
set of deformable object manipulation tasks and demonstrate that our method can
achieve effective and accurate servo-control for general deformable objects
with a wide variety of goal settings.Comment: arXiv admin note: text overlap with arXiv:1709.07218,
arXiv:1710.06947, arXiv:1802.0966
Human-Robot Collaboration: From Psychology to Social Robotics
With the advances in robotic technology, research in human-robot
collaboration (HRC) has gained in importance. For robots to interact with
humans autonomously they need active decision making that takes human partners
into account. However, state-of-the-art research in HRC does often assume a
leader-follower division, in which one agent leads the interaction. We believe
that this is caused by the lack of a reliable representation of the human and
the environment to allow autonomous decision making. This problem can be
overcome by an embodied approach to HRC which is inspired by psychological
studies of human-human interaction (HHI). In this survey, we review
neuroscientific and psychological findings of the sensorimotor patterns that
govern HHI and view them in a robotics context. Additionally, we study the
advances made by the robotic community into the direction of embodied HRC. We
focus on the mechanisms that are required for active, physical human-robot
collaboration. Finally, we discuss the similarities and differences in the two
fields of study which pinpoint directions of future research
Compare Contact Model-based Control and Contact Model-free Learning: A Survey of Robotic Peg-in-hole Assembly Strategies
In this paper, we present an overview of robotic peg-in-hole assembly and
analyze two main strategies: contact model-based and contact model-free
strategies. More specifically, we first introduce the contact model control
approaches, including contact state recognition and compliant control two
steps. Additionally, we focus on a comprehensive analysis of the whole robotic
assembly system. Second, without the contact state recognition process, we
decompose the contact model-free learning algorithms into two main subfields:
learning from demonstrations and learning from environments (mainly based on
reinforcement learning). For each subfield, we survey the landmark studies and
ongoing research to compare the different categories. We hope to strengthen the
relation between these two research communities by revealing the underlying
links. Ultimately, the remaining challenges and open questions in the field of
robotic peg-in-hole assembly community is discussed. The promising directions
and potential future work are also considered
Auto-conditioned Recurrent Mixture Density Networks for Learning Generalizable Robot Skills
Personal robots assisting humans must perform complex manipulation tasks that
are typically difficult to specify in traditional motion planning pipelines,
where multiple objectives must be met and the high-level context be taken into
consideration. Learning from demonstration (LfD) provides a promising way to
learn these kind of complex manipulation skills even from non-technical users.
However, it is challenging for existing LfD methods to efficiently learn skills
that can generalize to task specifications that are not covered by
demonstrations. In this paper, we introduce a state transition model (STM) that
generates joint-space trajectories by imitating motions from expert behavior.
Given a few demonstrations, we show in real robot experiments that the learned
STM can quickly generalize to unseen tasks and synthesize motions having longer
time horizons than the expert trajectories. Compared to conventional motion
planners, our approach enables the robot to accomplish complex behaviors from
high-level instructions without laborious hand-engineering of planning
objectives, while being able to adapt to changing goals during the skill
execution. In conjunction with a trajectory optimizer, our STM can construct a
high-quality skeleton of a trajectory that can be further improved in
smoothness and precision. In combination with a learned inverse dynamics model,
we additionally present results where the STM is used as a high-level planner.
A video of our experiments is available at https://youtu.be/85DX9Ojq-90Comment: Submitted to IROS 201
Pixels to Plans: Learning Non-Prehensile Manipulation by Imitating a Planner
We present a novel method enabling robots to quickly learn to manipulate
objects by leveraging a motion planner to generate "expert" training
trajectories from a small amount of human-labeled data. In contrast to the
traditional sense-plan-act cycle, we propose a deep learning architecture and
training regimen called PtPNet that can estimate effective end-effector
trajectories for manipulation directly from a single RGB-D image of an object.
Additionally, we present a data collection and augmentation pipeline that
enables the automatic generation of large numbers (millions) of training image
and trajectory examples with almost no human labeling effort.
We demonstrate our approach in a non-prehensile tool-based manipulation task,
specifically picking up shoes with a hook. In hardware experiments, PtPNet
generates motion plans (open-loop trajectories) that reliably (89% success over
189 trials) pick up four very different shoes from a range of positions and
orientations, and reliably picks up a shoe it has never seen before. Compared
with a traditional sense-plan-act paradigm, our system has the advantages of
operating on sparse information (single RGB-D frame), producing high-quality
trajectories much faster than the "expert" planner (300ms versus several
seconds), and generalizing effectively to previously unseen shoes.Comment: 8 page
Learning Movement Assessment Primitives for Force Interaction Skills
We present a novel, reusable and task-agnostic primitive for assessing the
outcome of a force-interaction robotic skill, useful e.g.\ for applications
such as quality control in industrial manufacturing. The proposed method is
easily programmed by kinesthetic teaching, and the desired adaptability and
reusability are achieved by machine learning models. The primitive records
sensory data during both demonstrations and reproductions of a movement.
Recordings include the end-effector's Cartesian pose and exerted wrench at each
time step. The collected data are then used to train Gaussian Processes which
create models of the wrench as a function of the robot's pose. The similarity
between the wrench models of the demonstration and the movement's reproduction
is derived by measuring their Hellinger distance. This comparison creates
features that are fed as inputs to a Naive Bayes classifier which estimates the
movement's probability of success. The evaluation is performed on two diverse
robotic assembly tasks -- snap-fitting and screwing -- with a total of 5 use
cases, 11 demonstrations, and more than 200 movement executions. The
performance metrics prove the proposed method's capability of generalization to
different demonstrations and movements
Vision-based Teleoperation of Shadow Dexterous Hand using End-to-End Deep Neural Network
In this paper, we present TeachNet, a novel neural network architecture for
intuitive and markerless vision-based teleoperation of dexterous robotic hands.
Robot joint angles are directly generated from depth images of the human hand
that produce visually similar robot hand poses in an end-to-end fashion. The
special structure of TeachNet, combined with a consistency loss function,
handles the differences in appearance and anatomy between human and robotic
hands. A synchronized human-robot training set is generated from an existing
dataset of labeled depth images of the human hand and simulated depth images of
a robotic hand. The final training set includes 400K pairwise depth images and
joint angles of a Shadow C6 robotic hand. The network evaluation results verify
the superiority of TeachNet, especially regarding the high-precision condition.
Imitation experiments and grasp tasks teleoperated by novice users demonstrate
that TeachNet is more reliable and faster than the state-of-the-art
vision-based teleoperation method.Comment: Accepted to ICRA 2019. Shuang Li and Xiaojian Ma contributed equally
to this wor
Autonomous Functional Locomotion in a Tendon-Driven Limb via Limited Experience
Robots will become ubiquitously useful only when they can use few attempts to
teach themselves to perform different tasks, even with complex bodies and in
dynamical environments. Vertebrates, in fact, successfully use trial-and-error
to learn multiple tasks in spite of their intricate tendon-driven anatomies.
Roboticists find such tendon-driven systems particularly hard to control
because they are simultaneously nonlinear, under-determined (many tendon
tensions combine to produce few net joint torques), and over-determined (few
joint rotations define how many tendons need to be reeled-in/payed-out). We
demonstrate---for the first time in simulation and in hardware---how a
model-free approach allows few-shot autonomous learning to produce effective
locomotion in a 3-tendon/2-joint tendon-driven leg. Initially, an artificial
neural network fed by sparsely sampled data collected using motor babbling
creates an inverse map from limb kinematics to motor activations, which is
analogous to juvenile vertebrates playing during development. Thereafter,
iterative reward-driven exploration of candidate motor activations
simultaneously refines the inverse map and finds a functional locomotor
limit-cycle autonomously. This biologically-inspired algorithm, which we call
G2P (General to Particular), enables versatile adaptation of robots to changes
in the target task, mechanics of their bodies, and environment. Moreover, this
work empowers future studies of few-shot autonomous learning in biological
systems, which is the foundation of their enviable functional versatility.Comment: 39 pages, 6 figure
Bayesian Disturbance Injection: Robust Imitation Learning of Flexible Policies
Scenarios requiring humans to choose from multiple seemingly optimal actions
are commonplace, however standard imitation learning often fails to capture
this behavior. Instead, an over-reliance on replicating expert actions induces
inflexible and unstable policies, leading to poor generalizability in an
application. To address the problem, this paper presents the first imitation
learning framework that incorporates Bayesian variational inference for
learning flexible non-parametric multi-action policies, while simultaneously
robustifying the policies against sources of error, by introducing and
optimizing disturbances to create a richer demonstration dataset. This
combinatorial approach forces the policy to adapt to challenging situations,
enabling stable multi-action policies to be learned efficiently. The
effectiveness of our proposed method is evaluated through simulations and
real-robot experiments for a table-sweep task using the UR3 6-DOF robotic arm.
Results show that, through improved flexibility and robustness, the learning
performance and control safety are better than comparison methods.Comment: 7 pages, Accepted by the 2021 International Conference on Robotics
and Automation (ICRA 2021
Guided Cost Learning: Deep Inverse Optimal Control via Policy Optimization
Reinforcement learning can acquire complex behaviors from high-level
specifications. However, defining a cost function that can be optimized
effectively and encodes the correct task is challenging in practice. We explore
how inverse optimal control (IOC) can be used to learn behaviors from
demonstrations, with applications to torque control of high-dimensional robotic
systems. Our method addresses two key challenges in inverse optimal control:
first, the need for informative features and effective regularization to impose
structure on the cost, and second, the difficulty of learning the cost function
under unknown dynamics for high-dimensional continuous systems. To address the
former challenge, we present an algorithm capable of learning arbitrary
nonlinear cost functions, such as neural networks, without meticulous feature
engineering. To address the latter challenge, we formulate an efficient
sample-based approximation for MaxEnt IOC. We evaluate our method on a series
of simulated tasks and real-world robotic manipulation problems, demonstrating
substantial improvement over prior methods both in terms of task complexity and
sample efficiency.Comment: International Conference on Machine Learning (ICML), 2016, to appea
- …