1,272 research outputs found
Combining Subgoal Graphs with Reinforcement Learning to Build a Rational Pathfinder
In this paper, we present a hierarchical path planning framework called SG-RL
(subgoal graphs-reinforcement learning), to plan rational paths for agents
maneuvering in continuous and uncertain environments. By "rational", we mean
(1) efficient path planning to eliminate first-move lags; (2) collision-free
and smooth for agents with kinematic constraints satisfied. SG-RL works in a
two-level manner. At the first level, SG-RL uses a geometric path-planning
method, i.e., Simple Subgoal Graphs (SSG), to efficiently find optimal abstract
paths, also called subgoal sequences. At the second level, SG-RL uses an RL
method, i.e., Least-Squares Policy Iteration (LSPI), to learn near-optimal
motion-planning policies which can generate kinematically feasible and
collision-free trajectories between adjacent subgoals. The first advantage of
the proposed method is that SSG can solve the limitations of sparse reward and
local minima trap for RL agents; thus, LSPI can be used to generate paths in
complex environments. The second advantage is that, when the environment
changes slightly (i.e., unexpected obstacles appearing), SG-RL does not need to
reconstruct subgoal graphs and replan subgoal sequences using SSG, since LSPI
can deal with uncertainties by exploiting its generalization ability to handle
changes in environments. Simulation experiments in representative scenarios
demonstrate that, compared with existing methods, SG-RL can work well on
large-scale maps with relatively low action-switching frequencies and shorter
path lengths, and SG-RL can deal with small changes in environments. We further
demonstrate that the design of reward functions and the types of training
environments are important factors for learning feasible policies.Comment: 20 page
Real-Time Planning with Multi-Fidelity Models for Agile Flights in Unknown Environments
Autonomous navigation through unknown environments is a challenging task that
entails real-time localization, perception, planning, and control. UAVs with
this capability have begun to emerge in the literature with advances in
lightweight sensing and computing. Although the planning methodologies vary
from platform to platform, many algorithms adopt a hierarchical planning
architecture where a slow, low-fidelity global planner guides a fast,
high-fidelity local planner. However, in unknown environments, this approach
can lead to erratic or unstable behavior due to the interaction between the
global planner, whose solution is changing constantly, and the local planner; a
consequence of not capturing higher-order dynamics in the global plan. This
work proposes a planning framework in which multi-fidelity models are used to
reduce the discrepancy between the local and global planner. Our approach uses
high-, medium-, and low-fidelity models to compose a path that captures
higher-order dynamics while remaining computationally tractable. In addition,
we address the interaction between a fast planner and a slower mapper by
considering the sensor data not yet fused into the map during the collision
check. This novel mapping and planning framework for agile flights is validated
in simulation and hardware experiments, showing replanning times of 5-40 ms in
cluttered environments.Comment: ICRA 201
Coverage and Time-optimal Motion Planning for Autonomous Vehicles
Autonomous vehicles are rapidly advancing with a variety of applications, such as area surveillance, environment mapping, and intelligent transportation. These applications require coverage and/or time-optimal motion planning, where the major challenges include uncertainties in the environment, motion constraints of vehicles, limited energy resources and potential failures. While dealing with these challenges in various capacities, this dissertation addresses three fundamental motion planning problems: (1) single-robot complete coverage in unknown environment, (2) multi-robot resilient and efficient coverage in unknown environment, and (3) time-optimal risk-aware motion planning for curvature-constrained vehicles.
First, the ε* algorithm is developed for online coverage path planning in unknown environment using a single autonomous vehicle. It is computationally efficient, and can generate the desired back-and-forth path with less turns and overlappings. ε* prevents the local extrema problem, thus can guarantee complete coverage. Second, the CARE algorithm is developed which extends ε* for multi-robot resilient and efficient coverage in unknown environment. In case of failures, CARE guarantees complete coverage via dynamic task reallocations of other vehicles, hence provides resilience. Moreover, it reallocates idling vehicles to support others in their tasks, hence improves efficiency. Finally, the T* algorithm is developed to find the time-optimal risk-aware path for curvature-constrained vehicles. We present a novel risk function based on the concept of collision time, and integrate it with the time cost for optimization.
The above-mentioned algorithms have been validated via simulations in complex scenarios and/or real experiments, and the results have shown clear advantages over existing popular approaches
Neural Sensor Fusion for Spatial Visualization on a Mobile Robot
An ARTMAP neural network is used to integrate visual information and ultrasonic sensory information on a B 14 mobile robot. Training samples for the neural network are acquired without human intervention. Sensory snapshots are retrospectively associated with the distance to the wall, provided by on~ board odomctry as the robot travels in a straight line. The goal is to produce a more accurate measure of distance than is provided by the raw sensors. The neural network effectively combines sensory sources both within and between modalities. The improved distance percept is used to produce occupancy grid visualizations of the robot's environment. The maps produced point to specific problems of raw sensory information processing and demonstrate the benefits of using a neural network system for sensor fusion.Office of Naval Research and Naval Research Laboratory (00014-96-1-0772, 00014-95-1-0409, 00014-95-0657
- …