1,123 research outputs found
Robot learning from demonstration of force-based manipulation tasks
One of the main challenges in Robotics is to develop robots that can interact with humans in a natural way, sharing the same dynamic and unstructured environments. Such an interaction may be aimed at assisting, helping or collaborating with a human user. To achieve this, the robot must be endowed with a cognitive system that allows it not only to learn new skills from its human partner, but also to refine or improve those already learned.
In this context, learning from demonstration appears as a natural and userfriendly way to transfer knowledge from humans to robots. This dissertation addresses such a topic and its application to an unexplored field, namely force-based manipulation tasks learning. In this kind of scenarios, force signals can convey data about the stiffness of a given object, the inertial components acting on a tool, a desired force profile to be reached, etc. Therefore, if the user wants the robot to learn a manipulation skill successfully, it is essential that its cognitive system is able to deal with force perceptions.
The first issue this thesis tackles is to extract the input information that is relevant for learning the task at hand, which is also known as the what to imitate? problem. Here, the proposed solution takes into consideration that the robot actions are a function of sensory signals, in other words the importance of each perception is assessed through its correlation with the robot movements. A Mutual Information analysis is used for selecting the most relevant inputs according to their influence on the output space. In this way, the robot can gather all the information coming from its sensory system, and the perception selection module proposed here automatically chooses the data the robot needs to learn a given task. Having selected the relevant input information for the task, it is necessary to represent the human demonstrations in a compact way, encoding the relevant characteristics of the data, for instance, sequential information, uncertainty, constraints, etc. This issue is the next problem addressed in this thesis. Here, a probabilistic learning framework based on hidden Markov models and Gaussian mixture regression is proposed for learning force-based manipulation skills. The outstanding features of such a framework are: (i) it is able to deal with the noise and uncertainty of force signals because of its probabilistic formulation, (ii) it exploits the sequential information embedded in the model for managing perceptual aliasing and time discrepancies, and (iii) it takes advantage of task variables to encode those force-based skills where the robot actions are modulated by an external parameter. Therefore, the resulting learning structure is able to robustly encode and reproduce different manipulation tasks.
After, this thesis goes a step forward by proposing a novel whole framework for learning impedance-based behaviors from demonstrations. The key aspects here are that this new structure merges vision and force information for encoding the data compactly, and it allows the robot to have different behaviors by shaping its compliance level over the course of the task. This is achieved by a parametric probabilistic model, whose Gaussian components are the basis of a statistical dynamical system that governs the robot motion.
From the force perceptions, the stiffness of the springs composing such a system are estimated, allowing the robot to shape its compliance. This approach permits to extend the learning paradigm to other fields different from the common trajectory following. The proposed frameworks are tested in three scenarios, namely, (a) the ball-in-box task, (b) drink pouring, and (c) a collaborative assembly, where the experimental results evidence the importance of using force perceptions as well as the usefulness and strengths of the methods
Geometry-aware Manipulability Learning, Tracking and Transfer
Body posture influences human and robots performance in manipulation tasks,
as appropriate poses facilitate motion or force exertion along different axes.
In robotics, manipulability ellipsoids arise as a powerful descriptor to
analyze, control and design the robot dexterity as a function of the
articulatory joint configuration. This descriptor can be designed according to
different task requirements, such as tracking a desired position or apply a
specific force. In this context, this paper presents a novel
\emph{manipulability transfer} framework, a method that allows robots to learn
and reproduce manipulability ellipsoids from expert demonstrations. The
proposed learning scheme is built on a tensor-based formulation of a Gaussian
mixture model that takes into account that manipulability ellipsoids lie on the
manifold of symmetric positive definite matrices. Learning is coupled with a
geometry-aware tracking controller allowing robots to follow a desired profile
of manipulability ellipsoids. Extensive evaluations in simulation with
redundant manipulators, a robotic hand and humanoids agents, as well as an
experiment with two real dual-arm systems validate the feasibility of the
approach.Comment: Accepted for publication in the Intl. Journal of Robotics Research
(IJRR). Website: https://sites.google.com/view/manipulability. Code:
https://github.com/NoemieJaquier/Manipulability. 24 pages, 20 figures, 3
tables, 4 appendice
A survey of robot manipulation in contact
In this survey, we present the current status on robots performing manipulation tasks that require varying contact with the environment, such that the robot must either implicitly or explicitly control the contact force with the environment to complete the task. Robots can perform more and more manipulation tasks that are still done by humans, and there is a growing number of publications on the topics of (1) performing tasks that always require contact and (2) mitigating uncertainty by leveraging the environment in tasks that, under perfect information, could be performed without contact. The recent trends have seen robots perform tasks earlier left for humans, such as massage, and in the classical tasks, such as peg-in-hole, there is a more efficient generalization to other similar tasks, better error tolerance, and faster planning or learning of the tasks. Thus, in this survey we cover the current stage of robots performing such tasks, starting from surveying all the different in-contact tasks robots can perform, observing how these tasks are controlled and represented, and finally presenting the learning and planning of the skills required to complete these tasks
Risk-sensitive Inverse Reinforcement Learning via Semi- and Non-Parametric Methods
The literature on Inverse Reinforcement Learning (IRL) typically assumes that
humans take actions in order to minimize the expected value of a cost function,
i.e., that humans are risk neutral. Yet, in practice, humans are often far from
being risk neutral. To fill this gap, the objective of this paper is to devise
a framework for risk-sensitive IRL in order to explicitly account for a human's
risk sensitivity. To this end, we propose a flexible class of models based on
coherent risk measures, which allow us to capture an entire spectrum of risk
preferences from risk-neutral to worst-case. We propose efficient
non-parametric algorithms based on linear programming and semi-parametric
algorithms based on maximum likelihood for inferring a human's underlying risk
measure and cost function for a rich class of static and dynamic
decision-making settings. The resulting approach is demonstrated on a simulated
driving game with ten human participants. Our method is able to infer and mimic
a wide range of qualitatively different driving styles from highly risk-averse
to risk-neutral in a data-efficient manner. Moreover, comparisons of the
Risk-Sensitive (RS) IRL approach with a risk-neutral model show that the RS-IRL
framework more accurately captures observed participant behavior both
qualitatively and quantitatively, especially in scenarios where catastrophic
outcomes such as collisions can occur.Comment: Submitted to International Journal of Robotics Research; Revision 1:
(i) Clarified minor technical points; (ii) Revised proof for Theorem 3 to
hold under weaker assumptions; (iii) Added additional figures and expanded
discussions to improve readabilit
Task Generalization with Stability Guarantees via Elastic Dynamical System Motion Policies
Dynamical System (DS) based Learning from Demonstration (LfD) allows learning
of reactive motion policies with stability and convergence guarantees from a
few trajectories. Yet, current DS learning techniques lack the flexibility to
generalize to new task instances as they ignore explicit task parameters that
inherently change the underlying trajectories. In this work, we propose
Elastic-DS, a novel DS learning, and generalization approach that embeds task
parameters into the Gaussian Mixture Model (GMM) based Linear Parameter Varying
(LPV) DS formulation. Central to our approach is the Elastic-GMM, a GMM
constrained to SE(3) task-relevant frames. Given a new task instance/context,
the Elastic-GMM is transformed with Laplacian Editing and used to re-estimate
the LPV-DS policy. Elastic-DS is compositional in nature and can be used to
construct flexible multi-step tasks. We showcase its strength on a myriad of
simulated and real-robot experiments while preserving desirable
control-theoretic guarantees. Supplementary videos can be found at
https://sites.google.com/view/elastic-dsComment: Accepted to CoRL 202
Comparative evaluation of approaches in T.4.1-4.3 and working definition of adaptive module
The goal of this deliverable is two-fold: (1) to present and compare different approaches towards learning and encoding movements us- ing dynamical systems that have been developed by the AMARSi partners (in the past during the first 6 months of the project), and (2) to analyze their suitability to be used as adaptive modules, i.e. as building blocks for the complete architecture that will be devel- oped in the project. The document presents a total of eight approaches, in two groups: modules for discrete movements (i.e. with a clear goal where the movement stops) and for rhythmic movements (i.e. which exhibit periodicity). The basic formulation of each approach is presented together with some illustrative simulation results. Key character- istics such as the type of dynamical behavior, learning algorithm, generalization properties, stability analysis are then discussed for each approach. We then make a comparative analysis of the different approaches by comparing these characteristics and discussing their suitability for the AMARSi project
Assisted Teleoperation in Changing Environments with a Mixture of Virtual Guides
Haptic guidance is a powerful technique to combine the strengths of humans
and autonomous systems for teleoperation. The autonomous system can provide
haptic cues to enable the operator to perform precise movements; the operator
can interfere with the plan of the autonomous system leveraging his/her
superior cognitive capabilities. However, providing haptic cues such that the
individual strengths are not impaired is challenging because low forces provide
little guidance, whereas strong forces can hinder the operator in realizing
his/her plan. Based on variational inference, we learn a Gaussian mixture model
(GMM) over trajectories to accomplish a given task. The learned GMM is used to
construct a potential field which determines the haptic cues. The potential
field smoothly changes during teleoperation based on our updated belief over
the plans and their respective phases. Furthermore, new plans are learned
online when the operator does not follow any of the proposed plans, or after
changes in the environment. User studies confirm that our framework helps users
perform teleoperation tasks more accurately than without haptic cues and, in
some cases, faster. Moreover, we demonstrate the use of our framework to help a
subject teleoperate a 7 DoF manipulator in a pick-and-place task.Comment: 19 pages, 9 figure
- …