29,727 research outputs found
Static and Dynamic Path Planning Using Incremental Heuristic Search
Path planning is an important component in any highly automated vehicle
system. In this report, the general problem of path planning is considered
first in partially known static environments where only static obstacles are
present but the layout of the environment is changing as the agent acquires new
information. Attention is then given to the problem of path planning in dynamic
environments where there are moving obstacles in addition to the static ones.
Specifically, a 2D car-like agent traversing in a 2D environment was
considered. It was found that the traditional configuration-time space approach
is unsuitable for producing trajectories consistent with the dynamic
constraints of a car. A novel scheme is then suggested where the state space is
4D consisting of position, speed and time but the search is done in the 3D
space composed by position and speed. Simulation tests shows that the new
scheme is capable of efficiently producing trajectories respecting the dynamic
constraint of a car-like agent with a bound on their optimality.Comment: Internship Repor
Real-Time Planning with Multi-Fidelity Models for Agile Flights in Unknown Environments
Autonomous navigation through unknown environments is a challenging task that
entails real-time localization, perception, planning, and control. UAVs with
this capability have begun to emerge in the literature with advances in
lightweight sensing and computing. Although the planning methodologies vary
from platform to platform, many algorithms adopt a hierarchical planning
architecture where a slow, low-fidelity global planner guides a fast,
high-fidelity local planner. However, in unknown environments, this approach
can lead to erratic or unstable behavior due to the interaction between the
global planner, whose solution is changing constantly, and the local planner; a
consequence of not capturing higher-order dynamics in the global plan. This
work proposes a planning framework in which multi-fidelity models are used to
reduce the discrepancy between the local and global planner. Our approach uses
high-, medium-, and low-fidelity models to compose a path that captures
higher-order dynamics while remaining computationally tractable. In addition,
we address the interaction between a fast planner and a slower mapper by
considering the sensor data not yet fused into the map during the collision
check. This novel mapping and planning framework for agile flights is validated
in simulation and hardware experiments, showing replanning times of 5-40 ms in
cluttered environments.Comment: ICRA 201
Intrinsic Motivation and Mental Replay enable Efficient Online Adaptation in Stochastic Recurrent Networks
Autonomous robots need to interact with unknown, unstructured and changing
environments, constantly facing novel challenges. Therefore, continuous online
adaptation for lifelong-learning and the need of sample-efficient mechanisms to
adapt to changes in the environment, the constraints, the tasks, or the robot
itself are crucial. In this work, we propose a novel framework for
probabilistic online motion planning with online adaptation based on a
bio-inspired stochastic recurrent neural network. By using learning signals
which mimic the intrinsic motivation signalcognitive dissonance in addition
with a mental replay strategy to intensify experiences, the stochastic
recurrent network can learn from few physical interactions and adapts to novel
environments in seconds. We evaluate our online planning and adaptation
framework on an anthropomorphic KUKA LWR arm. The rapid online adaptation is
shown by learning unknown workspace constraints sample-efficiently from few
physical interactions while following given way points.Comment: accepted in Neural Network
Combining Subgoal Graphs with Reinforcement Learning to Build a Rational Pathfinder
In this paper, we present a hierarchical path planning framework called SG-RL
(subgoal graphs-reinforcement learning), to plan rational paths for agents
maneuvering in continuous and uncertain environments. By "rational", we mean
(1) efficient path planning to eliminate first-move lags; (2) collision-free
and smooth for agents with kinematic constraints satisfied. SG-RL works in a
two-level manner. At the first level, SG-RL uses a geometric path-planning
method, i.e., Simple Subgoal Graphs (SSG), to efficiently find optimal abstract
paths, also called subgoal sequences. At the second level, SG-RL uses an RL
method, i.e., Least-Squares Policy Iteration (LSPI), to learn near-optimal
motion-planning policies which can generate kinematically feasible and
collision-free trajectories between adjacent subgoals. The first advantage of
the proposed method is that SSG can solve the limitations of sparse reward and
local minima trap for RL agents; thus, LSPI can be used to generate paths in
complex environments. The second advantage is that, when the environment
changes slightly (i.e., unexpected obstacles appearing), SG-RL does not need to
reconstruct subgoal graphs and replan subgoal sequences using SSG, since LSPI
can deal with uncertainties by exploiting its generalization ability to handle
changes in environments. Simulation experiments in representative scenarios
demonstrate that, compared with existing methods, SG-RL can work well on
large-scale maps with relatively low action-switching frequencies and shorter
path lengths, and SG-RL can deal with small changes in environments. We further
demonstrate that the design of reward functions and the types of training
environments are important factors for learning feasible policies.Comment: 20 page
- …