533 research outputs found
Provably Safe Robot Navigation with Obstacle Uncertainty
As drones and autonomous cars become more widespread it is becoming
increasingly important that robots can operate safely under realistic
conditions. The noisy information fed into real systems means that robots must
use estimates of the environment to plan navigation. Efficiently guaranteeing
that the resulting motion plans are safe under these circumstances has proved
difficult. We examine how to guarantee that a trajectory or policy is safe with
only imperfect observations of the environment. We examine the implications of
various mathematical formalisms of safety and arrive at a mathematical notion
of safety of a long-term execution, even when conditioned on observational
information. We present efficient algorithms that can prove that trajectories
or policies are safe with much tighter bounds than in previous work. Notably,
the complexity of the environment does not affect our methods ability to
evaluate if a trajectory or policy is safe. We then use these safety checking
methods to design a safe variant of the RRT planning algorithm.Comment: RSS 201
Deep Network Uncertainty Maps for Indoor Navigation
Most mobile robots for indoor use rely on 2D laser scanners for localization,
mapping and navigation. These sensors, however, cannot detect transparent
surfaces or measure the full occupancy of complex objects such as tables. Deep
Neural Networks have recently been proposed to overcome this limitation by
learning to estimate object occupancy. These estimates are nevertheless subject
to uncertainty, making the evaluation of their confidence an important issue
for these measures to be useful for autonomous navigation and mapping. In this
work we approach the problem from two sides. First we discuss uncertainty
estimation in deep models, proposing a solution based on a fully convolutional
neural network. The proposed architecture is not restricted by the assumption
that the uncertainty follows a Gaussian model, as in the case of many popular
solutions for deep model uncertainty estimation, such as Monte-Carlo Dropout.
We present results showing that uncertainty over obstacle distances is actually
better modeled with a Laplace distribution. Then, we propose a novel approach
to build maps based on Deep Neural Network uncertainty models. In particular,
we present an algorithm to build a map that includes information over obstacle
distance estimates while taking into account the level of uncertainty in each
estimate. We show how the constructed map can be used to increase global
navigation safety by planning trajectories which avoid areas of high
uncertainty, enabling higher autonomy for mobile robots in indoor settings.Comment: Accepted for publication in "2019 IEEE-RAS International Conference
on Humanoid Robots (Humanoids)
SAFETY-GUARANTEED TASK PLANNING FOR BIPEDAL NAVIGATION IN PARTIALLY OBSERVABLE ENVIRONMENTS
Bipedal robots are becoming more capable as basic hardware and control challenges are being overcome, however reasoning about safety at the task and motion planning levels has been largely underexplored. I would like to make key steps towards guaranteeing safe locomotion in cluttered environments in the presence of humans or other dynamic obstacles by designing a hierarchical task planning framework that incorporates safety guarantees at each level. This layered planning framework is composed of a coarse high-level symbolic navigation planner and a lower-level local action planner. A belief abstraction at the global navigation planning level enables belief estimation of non-visible dynamic obstacle states and guarantees navigation safety with collision avoidance. Both planning layers employ linear temporal logic for a reactive game synthesis between the robot and its environment while incorporating lower level safe locomotion keyframe policies into formal task specification design.
The high-level symbolic navigation planner has been extended to leverage the capabilities of a heterogeneous multi-agent team to resolve environment assumption violations that appear at runtime. Modifications in the navigation planner in conjunction with a coordination layer allow each agent to guarantee immediate safety and eventual task completion in the presence of an assumption violation if another agent exists that can resolve said violation, e.g. a door is closed that another dexterous agent can open.
The planning framework leverages the expressive nature and formal guarantees of LTL to generate provably correct controllers for complex robotic systems. The use of belief space planning for dynamic obstacle belief tracking and heterogeneous robot capabilities to assist one another when environment assumptions are violated allows the planning framework to reduce the conservativeness traditionally associated with using formal methods for robot planning.M.S
- …