149 research outputs found
Feedback-based Fabric Strip Folding
Accurate manipulation of a deformable body such as a piece of fabric is
difficult because of its many degrees of freedom and unobservable properties
affecting its dynamics. To alleviate these challenges, we propose the
application of feedback-based control to robotic fabric strip folding. The
feedback is computed from the low dimensional state extracted from a camera
image. We trained the controller using reinforcement learning in simulation
which was calibrated to cover the real fabric strip behaviors. The proposed
feedback-based folding was experimentally compared to two state-of-the-art
folding methods and our method outperformed both of them in terms of accuracy.Comment: Submitted to IEEE/RSJ IROS201
Deep Network Uncertainty Maps for Indoor Navigation
Most mobile robots for indoor use rely on 2D laser scanners for localization,
mapping and navigation. These sensors, however, cannot detect transparent
surfaces or measure the full occupancy of complex objects such as tables. Deep
Neural Networks have recently been proposed to overcome this limitation by
learning to estimate object occupancy. These estimates are nevertheless subject
to uncertainty, making the evaluation of their confidence an important issue
for these measures to be useful for autonomous navigation and mapping. In this
work we approach the problem from two sides. First we discuss uncertainty
estimation in deep models, proposing a solution based on a fully convolutional
neural network. The proposed architecture is not restricted by the assumption
that the uncertainty follows a Gaussian model, as in the case of many popular
solutions for deep model uncertainty estimation, such as Monte-Carlo Dropout.
We present results showing that uncertainty over obstacle distances is actually
better modeled with a Laplace distribution. Then, we propose a novel approach
to build maps based on Deep Neural Network uncertainty models. In particular,
we present an algorithm to build a map that includes information over obstacle
distance estimates while taking into account the level of uncertainty in each
estimate. We show how the constructed map can be used to increase global
navigation safety by planning trajectories which avoid areas of high
uncertainty, enabling higher autonomy for mobile robots in indoor settings.Comment: Accepted for publication in "2019 IEEE-RAS International Conference
on Humanoid Robots (Humanoids)
Differential Dynamic Programming with Nonlinear Safety Constraints Under System Uncertainties
Safe operation of systems such as robots requires them to plan and execute
trajectories subject to safety constraints. When those systems are subject to
uncertainties in their dynamics, ensuring that the constraints are not violated
is challenging. In this paper, we propose a safe trajectory optimization and
control approach (Safe-CDDP) for systems under additive uncertainties and
non-linear safety constraints based on constrained differential dynamic
programming (DDP). The safety of the robot during its motion is formulated as
chance-constraints with user-chosen probabilities of constraint satisfaction.
The chance constraints are transformed into deterministic ones in DDP
formulation by constraint tightening. To avoid over conservatism during
constraint tightening, linear control gains of the feedback policy derived from
the constrained DDP are used in the approximation of closed-loop uncertainty
propagation in prediction. The proposed algorithm is empirically demonstrated
on three different robot dynamics with up to 12 states and the results show the
applicability of the approach for safety-aware applications.Comment: 7 pages, 4 figures, submitted to ICRA 202
Increasing the Efficiency of Policy Learning for Autonomous Vehicles by Multi-Task Representation Learning
Driving in a dynamic, multi-agent, and complex urban environment is a
difficult task requiring a complex decision-making policy. The learning of such
a policy requires a state representation that can encode the entire
environment. Mid-level representations that encode a vehicle's environment as
images have become a popular choice. Still, they are quite high-dimensional,
limiting their use in data-hungry approaches such as reinforcement learning. In
this article, we propose to learn a low-dimensional and rich latent
representation of the environment by leveraging the knowledge of relevant
semantic factors. To do this, we train an encoder-decoder deep neural network
to predict multiple application-relevant factors such as the trajectories of
other agents and the ego car. Furthermore, we propose a hazard signal based on
other vehicles' future trajectories and the planned route which is used in
conjunction with the learned latent representation as input to a down-stream
policy. We demonstrate that using the multi-head encoder-decoder neural network
results in a more informative representation than a standard single-head model.
In particular, the proposed representation learning and the hazard signal help
reinforcement learning to learn faster, with increased performance and less
data than baseline methods
Robotic manipulation of multiple objects as a POMDP
This paper investigates manipulation of multiple unknown objects in a crowded
environment. Because of incomplete knowledge due to unknown objects and
occlusions in visual observations, object observations are imperfect and action
success is uncertain, making planning challenging. We model the problem as a
partially observable Markov decision process (POMDP), which allows a general
reward based optimization objective and takes uncertainty in temporal evolution
and partial observations into account. In addition to occlusion dependent
observation and action success probabilities, our POMDP model also
automatically adapts object specific action success probabilities. To cope with
the changing system dynamics and performance constraints, we present a new
online POMDP method based on particle filtering that produces compact policies.
The approach is validated both in simulation and in physical experiments in a
scenario of moving dirty dishes into a dishwasher. The results indicate that:
1) a greedy heuristic manipulation approach is not sufficient, multi-object
manipulation requires multi-step POMDP planning, and 2) on-line planning is
beneficial since it allows the adaptation of the system dynamics model based on
actual experience
- …