1,995 research outputs found
Risk-sensitive Inverse Reinforcement Learning via Semi- and Non-Parametric Methods
The literature on Inverse Reinforcement Learning (IRL) typically assumes that
humans take actions in order to minimize the expected value of a cost function,
i.e., that humans are risk neutral. Yet, in practice, humans are often far from
being risk neutral. To fill this gap, the objective of this paper is to devise
a framework for risk-sensitive IRL in order to explicitly account for a human's
risk sensitivity. To this end, we propose a flexible class of models based on
coherent risk measures, which allow us to capture an entire spectrum of risk
preferences from risk-neutral to worst-case. We propose efficient
non-parametric algorithms based on linear programming and semi-parametric
algorithms based on maximum likelihood for inferring a human's underlying risk
measure and cost function for a rich class of static and dynamic
decision-making settings. The resulting approach is demonstrated on a simulated
driving game with ten human participants. Our method is able to infer and mimic
a wide range of qualitatively different driving styles from highly risk-averse
to risk-neutral in a data-efficient manner. Moreover, comparisons of the
Risk-Sensitive (RS) IRL approach with a risk-neutral model show that the RS-IRL
framework more accurately captures observed participant behavior both
qualitatively and quantitatively, especially in scenarios where catastrophic
outcomes such as collisions can occur.Comment: Submitted to International Journal of Robotics Research; Revision 1:
(i) Clarified minor technical points; (ii) Revised proof for Theorem 3 to
hold under weaker assumptions; (iii) Added additional figures and expanded
discussions to improve readabilit
Optimization Model for Planning Precision Grasps with Multi-Fingered Hands
Precision grasps with multi-fingered hands are important for precise
placement and in-hand manipulation tasks. Searching precision grasps on the
object represented by point cloud, is challenging due to the complex object
shape, high-dimensionality, collision and undesired properties of the sensing
and positioning. This paper proposes an optimization model to search for
precision grasps with multi-fingered hands. The model takes noisy point cloud
of the object as input and optimizes the grasp quality by iteratively searching
for the palm pose and finger joints positions. The collision between the hand
and the object is approximated and penalized by a series of least-squares. The
collision approximation is able to handle the point cloud representation of the
objects with complex shapes. The proposed optimization model is able to locate
collision-free optimal precision grasps efficiently. The average computation
time is 0.50 sec/grasp. The searching is robust to the incompleteness and noise
of the point cloud. The effectiveness of the algorithm is demonstrated by
experiments.Comment: Submitted to IROS2019, experiment on BarrettHand, 8 page
Learning Task Specifications from Demonstrations
Real world applications often naturally decompose into several sub-tasks. In
many settings (e.g., robotics) demonstrations provide a natural way to specify
the sub-tasks. However, most methods for learning from demonstrations either do
not provide guarantees that the artifacts learned for the sub-tasks can be
safely recombined or limit the types of composition available. Motivated by
this deficit, we consider the problem of inferring Boolean non-Markovian
rewards (also known as logical trace properties or specifications) from
demonstrations provided by an agent operating in an uncertain, stochastic
environment. Crucially, specifications admit well-defined composition rules
that are typically easy to interpret. In this paper, we formulate the
specification inference task as a maximum a posteriori (MAP) probability
inference problem, apply the principle of maximum entropy to derive an analytic
demonstration likelihood model and give an efficient approach to search for the
most likely specification in a large candidate pool of specifications. In our
experiments, we demonstrate how learning specifications can help avoid common
problems that often arise due to ad-hoc reward composition.Comment: NIPS 201
Mixture of Interaction Primitives for Multiple Agents
abstract: In a collaborative environment where multiple robots and human beings are expected
to collaborate to perform a task, it becomes essential for a robot to be aware of multiple
agents working in its work environment. A robot must also learn to adapt to
different agents in the workspace and conduct its interaction based on the presence
of these agents. A theoretical framework was introduced which performs interaction
learning from demonstrations in a two-agent work environment, and it is called
Interaction Primitives.
This document is an in-depth description of the new state of the art Python
Framework for Interaction Primitives between two agents in a single as well as multiple
task work environment and extension of the original framework in a work environment
with multiple agents doing a single task. The original theory of Interaction
Primitives has been extended to create a framework which will capture correlation
between more than two agents while performing a single task. The new state of the
art Python framework is an intuitive, generic, easy to install and easy to use python
library which can be applied to use the Interaction Primitives framework in a work
environment. This library was tested in simulated environments and controlled laboratory
environment. The results and benchmarks of this library are available in the
related sections of this document.Dissertation/ThesisMasters Thesis Computer Science 201
Robotic learning of force-based industrial manipulation tasks
Even with the rapid technological advancements, robots are still not the most comfortable machines to work with. Firstly, due to the separation of the robot and human workspace which imposes an additional financial burden. Secondly, due to the significant re-programming cost in case of changing products, especially in Small and Medium-sized Enterprises (SMEs). Therefore, there is a significant need to reduce the programming efforts required to enable robots to perform various tasks while sharing the same space with a human operator. Hence, the robot must be equipped with a cognitive and perceptual capabilities that facilitate human-robot interaction. Humans use their various senses to perform tasks such as vision, smell and taste. One sensethat plays a significant role in human activity is ’touch’ or ’force’. For example, holding a cup of tea, or making fine adjustments while inserting a key requires haptic information to achieve the task successfully. In all these examples, force and torque data are crucial for the successful completion of the activity. Also, this information implicitly conveys data about contact force, object stiffness, and many others. Hence, a deep understanding of the execution of such events can bridge the gap between humans and robots. This thesis is being directed to equip an industrial robot with the ability to deal with force perceptions and then learn force-based tasks using Learning from Demonstration (LfD).To learn force-based tasks using LfD, it is essential to extract task-relevant features from the force information. Then, knowledge must be extracted and encoded form the task-relevant features. Hence, the captured skills can be reproduced in a new scenario. In this thesis, these elements of LfD were achieved using different approaches based on the demonstrated task. In this thesis, four robotics problems were addressed using LfD framework. The first challenge was to filter out robots’ internal forces (irrelevant signals) using data-driven approach. The second robotics challenge was the recognition of the Contact State (CS) during assembly tasks. To tackle this challenge, a symbolic based approach was proposed, in which a force/torque signals; during demonstrated assembly, the task was encoded as a sequence of symbols. The third challenge was to learn a human-robot co-manipulation task based on LfD. In this case, an ensemble machine learning approach was proposed to capture such a skill. The last challenge in this thesis, was to learn an assembly task by demonstration with the presents of parts geometrical variation. Hence, a new learning approach based on Artificial Potential Field (APF) to learn a Peg-in-Hole (PiH) assembly task which includes no-contact and contact phases. To sum up, this thesis focuses on the use of data-driven approaches to learning force based task in an industrial context. Hence, different machine learning approaches were implemented, developed and evaluated in different scenarios. Then, the performance of these approaches was compared with mathematical modelling based approaches.</div
- …