1,882 research outputs found
Finding a needle in an exponential haystack: Discrete RRT for exploration of implicit roadmaps in multi-robot motion planning
We present a sampling-based framework for multi-robot motion planning which
combines an implicit representation of a roadmap with a novel approach for
pathfinding in geometrically embedded graphs tailored for our setting. Our
pathfinding algorithm, discrete-RRT (dRRT), is an adaptation of the celebrated
RRT algorithm for the discrete case of a graph, and it enables a rapid
exploration of the high-dimensional configuration space by carefully walking
through an implicit representation of a tensor product of roadmaps for the
individual robots. We demonstrate our approach experimentally on scenarios of
up to 60 degrees of freedom where our algorithm is faster by a factor of at
least ten when compared to existing algorithms that we are aware of.Comment: Kiril Solovey and Oren Salzman contributed equally to this pape
PRM-RL: Long-range Robotic Navigation Tasks by Combining Reinforcement Learning and Sampling-based Planning
We present PRM-RL, a hierarchical method for long-range navigation task
completion that combines sampling based path planning with reinforcement
learning (RL). The RL agents learn short-range, point-to-point navigation
policies that capture robot dynamics and task constraints without knowledge of
the large-scale topology. Next, the sampling-based planners provide roadmaps
which connect robot configurations that can be successfully navigated by the RL
agent. The same RL agents are used to control the robot under the direction of
the planning, enabling long-range navigation. We use the Probabilistic Roadmaps
(PRMs) for the sampling-based planner. The RL agents are constructed using
feature-based and deep neural net policies in continuous state and action
spaces. We evaluate PRM-RL, both in simulation and on-robot, on two navigation
tasks with non-trivial robot dynamics: end-to-end differential drive indoor
navigation in office environments, and aerial cargo delivery in urban
environments with load displacement constraints. Our results show improvement
in task completion over both RL agents on their own and traditional
sampling-based planners. In the indoor navigation task, PRM-RL successfully
completes up to 215 m long trajectories under noisy sensor conditions, and the
aerial cargo delivery completes flights over 1000 m without violating the task
constraints in an environment 63 million times larger than used in training.Comment: 9 pages, 7 figure
Experience-Based Planning with Sparse Roadmap Spanners
We present an experienced-based planning framework called Thunder that learns
to reduce computation time required to solve high-dimensional planning problems
in varying environments. The approach is especially suited for large
configuration spaces that include many invariant constraints, such as those
found with whole body humanoid motion planning. Experiences are generated using
probabilistic sampling and stored in a sparse roadmap spanner (SPARS), which
provides asymptotically near-optimal coverage of the configuration space,
making storing, retrieving, and repairing past experiences very efficient with
respect to memory and time. The Thunder framework improves upon past
experience-based planners by storing experiences in a graph rather than in
individual paths, eliminating redundant information, providing more
opportunities for path reuse, and providing a theoretical limit to the size of
the experience graph. These properties also lead to improved handling of
dynamically changing environments, reasoning about optimal paths, and reducing
query resolution time. The approach is demonstrated on a 30 degrees of freedom
humanoid robot and compared with the Lightning framework, an experience-based
planner that uses individual paths to store past experiences. In environments
with variable obstacles and stability constraints, experiments show that
Thunder is on average an order of magnitude faster than Lightning and planning
from scratch. Thunder also uses 98.8% less memory to store its experiences
after 10,000 trials when compared to Lightning. Our framework is implemented
and freely available in the Open Motion Planning Library.Comment: Submitted to ICRA 201
- …