35 research outputs found
Experience-Based Planning with Sparse Roadmap Spanners
We present an experienced-based planning framework called Thunder that learns
to reduce computation time required to solve high-dimensional planning problems
in varying environments. The approach is especially suited for large
configuration spaces that include many invariant constraints, such as those
found with whole body humanoid motion planning. Experiences are generated using
probabilistic sampling and stored in a sparse roadmap spanner (SPARS), which
provides asymptotically near-optimal coverage of the configuration space,
making storing, retrieving, and repairing past experiences very efficient with
respect to memory and time. The Thunder framework improves upon past
experience-based planners by storing experiences in a graph rather than in
individual paths, eliminating redundant information, providing more
opportunities for path reuse, and providing a theoretical limit to the size of
the experience graph. These properties also lead to improved handling of
dynamically changing environments, reasoning about optimal paths, and reducing
query resolution time. The approach is demonstrated on a 30 degrees of freedom
humanoid robot and compared with the Lightning framework, an experience-based
planner that uses individual paths to store past experiences. In environments
with variable obstacles and stability constraints, experiments show that
Thunder is on average an order of magnitude faster than Lightning and planning
from scratch. Thunder also uses 98.8% less memory to store its experiences
after 10,000 trials when compared to Lightning. Our framework is implemented
and freely available in the Open Motion Planning Library.Comment: Submitted to ICRA 201
Learning How to Walk: Warm-starting Optimal Control Solver with Memory of Motion
In this paper, we propose a framework to build a memory of motion for warm-starting an optimal control solver for the locomotion task of a humanoid robot. We use HPP Loco3D, a versatile locomotion planner, to generate offline a set of dynamically consistent whole-body trajectory to be stored as the memory of motion. The learning problem is formulated as a regression problem to predict a single-step motion given the desired contact locations, which is used as a building block for producing multi-step motions. The predicted motion is then used as a warm-start for the fast optimal control solver Crocoddyl. We have shown that the approach manages to reduce the required number of iterations to reach the convergence from ~9.5 to only ~3.0 iterations for the single-step motion and from ~6.2 to ~4.5 iterations for the multi-step motion, while maintaining the solution's quality
Reliable Trajectories for Dynamic Quadrupeds using Analytical Costs and Learned Initializations
Dynamic traversal of uneven terrain is a major objective in the field of
legged robotics. The most recent model predictive control approaches for these
systems can generate robust dynamic motion of short duration; however, planning
over a longer time horizon may be necessary when navigating complex terrain. A
recently-developed framework, Trajectory Optimization for Walking Robots
(TOWR), computes such plans but does not guarantee their reliability on real
platforms, under uncertainty and perturbations. We extend TOWR with analytical
costs to generate trajectories that a state-of-the-art whole-body tracking
controller can successfully execute. To reduce online computation time, we
implement a learning-based scheme for initialization of the nonlinear program
based on offline experience. The execution of trajectories as long as 16
footsteps and 5.5 s over different terrains by a real quadruped demonstrates
the effectiveness of the approach on hardware. This work builds toward an
online system which can efficiently and robustly replan dynamic trajectories.Comment: Video: https://youtu.be/LKFDB_BOhl