57,397 research outputs found
Combining Subgoal Graphs with Reinforcement Learning to Build a Rational Pathfinder
In this paper, we present a hierarchical path planning framework called SG-RL
(subgoal graphs-reinforcement learning), to plan rational paths for agents
maneuvering in continuous and uncertain environments. By "rational", we mean
(1) efficient path planning to eliminate first-move lags; (2) collision-free
and smooth for agents with kinematic constraints satisfied. SG-RL works in a
two-level manner. At the first level, SG-RL uses a geometric path-planning
method, i.e., Simple Subgoal Graphs (SSG), to efficiently find optimal abstract
paths, also called subgoal sequences. At the second level, SG-RL uses an RL
method, i.e., Least-Squares Policy Iteration (LSPI), to learn near-optimal
motion-planning policies which can generate kinematically feasible and
collision-free trajectories between adjacent subgoals. The first advantage of
the proposed method is that SSG can solve the limitations of sparse reward and
local minima trap for RL agents; thus, LSPI can be used to generate paths in
complex environments. The second advantage is that, when the environment
changes slightly (i.e., unexpected obstacles appearing), SG-RL does not need to
reconstruct subgoal graphs and replan subgoal sequences using SSG, since LSPI
can deal with uncertainties by exploiting its generalization ability to handle
changes in environments. Simulation experiments in representative scenarios
demonstrate that, compared with existing methods, SG-RL can work well on
large-scale maps with relatively low action-switching frequencies and shorter
path lengths, and SG-RL can deal with small changes in environments. We further
demonstrate that the design of reward functions and the types of training
environments are important factors for learning feasible policies.Comment: 20 page
Search-based 3D Planning and Trajectory Optimization for Safe Micro Aerial Vehicle Flight Under Sensor Visibility Constraints
Safe navigation of Micro Aerial Vehicles (MAVs) requires not only
obstacle-free flight paths according to a static environment map, but also the
perception of and reaction to previously unknown and dynamic objects. This
implies that the onboard sensors cover the current flight direction. Due to the
limited payload of MAVs, full sensor coverage of the environment has to be
traded off with flight time. Thus, often only a part of the environment is
covered.
We present a combined allocentric complete planning and trajectory
optimization approach taking these sensor visibility constraints into account.
The optimized trajectories yield flight paths within the apex angle of a
Velodyne Puck Lite 3D laser scanner enabling low-level collision avoidance to
perceive obstacles in the flight direction. Furthermore, the optimized
trajectories take the flight dynamics into account and contain the velocities
and accelerations along the path.
We evaluate our approach with a DJI Matrice 600 MAV and in simulation
employing hardware-in-the-loop.Comment: In Proceedings of IEEE International Conference on Robotics and
Automation (ICRA), Montreal, Canada, May 201
Downwash-Aware Trajectory Planning for Large Quadrotor Teams
We describe a method for formation-change trajectory planning for large
quadrotor teams in obstacle-rich environments. Our method decomposes the
planning problem into two stages: a discrete planner operating on a graph
representation of the workspace, and a continuous refinement that converts the
non-smooth graph plan into a set of C^k-continuous trajectories, locally
optimizing an integral-squared-derivative cost. We account for the downwash
effect, allowing safe flight in dense formations. We demonstrate the
computational efficiency in simulation with up to 200 robots and the physical
plausibility with an experiment with 32 nano-quadrotors. Our approach can
compute safe and smooth trajectories for hundreds of quadrotors in dense
environments with obstacles in a few minutes.Comment: 8 page
Fast Manipulability Maximization Using Continuous-Time Trajectory Optimization
A significant challenge in manipulation motion planning is to ensure agility
in the face of unpredictable changes during task execution. This requires the
identification and possible modification of suitable joint-space trajectories,
since the joint velocities required to achieve a specific endeffector motion
vary with manipulator configuration. For a given manipulator configuration, the
joint space-to-task space velocity mapping is characterized by a quantity known
as the manipulability index. In contrast to previous control-based approaches,
we examine the maximization of manipulability during planning as a way of
achieving adaptable and safe joint space-to-task space motion mappings in
various scenarios. By representing the manipulator trajectory as a
continuous-time Gaussian process (GP), we are able to leverage recent advances
in trajectory optimization to maximize the manipulability index during
trajectory generation. Moreover, the sparsity of our chosen representation
reduces the typically large computational cost associated with maximizing
manipulability when additional constraints exist. Results from simulation
studies and experiments with a real manipulator demonstrate increases in
manipulability, while maintaining smooth trajectories with more dexterous (and
therefore more agile) arm configurations.Comment: In Proceedings of the IEEE International Conference on Intelligent
Robots and Systems (IROS'19), Macau, China, Nov. 4-8, 201
- …