13,352 research outputs found
Dynamic Motion Planning for Aerial Surveillance on a Fixed-Wing UAV
We present an efficient path planning algorithm for an Unmanned Aerial
Vehicle surveying a cluttered urban landscape. A special emphasis is on
maximizing area surveyed while adhering to constraints of the UAV and partially
known and updating environment. A Voronoi bias is introduced in the
probabilistic roadmap building phase to identify certain critical milestones
for maximal surveillance of the search space. A kinematically feasible but
coarse tour connecting these milestones is generated by the global path
planner. A local path planner then generates smooth motion primitives between
consecutive nodes of the global path based on UAV as a Dubins vehicle and
taking into account any impending obstacles. A Markov Decision Process (MDP)
models the control policy for the UAV and determines the optimal action to be
undertaken for evading the obstacles in the vicinity with minimal deviation
from current path. The efficacy of the proposed algorithm is evaluated in an
updating simulation environment with dynamic and static obstacles.Comment: Accepted at International Conference on Unmanned Aircraft Systems
201
Human Motion Trajectory Prediction: A Survey
With growing numbers of intelligent autonomous systems in human environments,
the ability of such systems to perceive, understand and anticipate human
behavior becomes increasingly important. Specifically, predicting future
positions of dynamic agents and planning considering such predictions are key
tasks for self-driving vehicles, service robots and advanced surveillance
systems. This paper provides a survey of human motion trajectory prediction. We
review, analyze and structure a large selection of work from different
communities and propose a taxonomy that categorizes existing methods based on
the motion modeling approach and level of contextual information used. We
provide an overview of the existing datasets and performance metrics. We
discuss limitations of the state of the art and outline directions for further
research.Comment: Submitted to the International Journal of Robotics Research (IJRR),
37 page
Imitating Driver Behavior with Generative Adversarial Networks
The ability to accurately predict and simulate human driving behavior is
critical for the development of intelligent transportation systems. Traditional
modeling methods have employed simple parametric models and behavioral cloning.
This paper adopts a method for overcoming the problem of cascading errors
inherent in prior approaches, resulting in realistic behavior that is robust to
trajectory perturbations. We extend Generative Adversarial Imitation Learning
to the training of recurrent policies, and we demonstrate that our model
outperforms rule-based controllers and maximum likelihood models in realistic
highway simulations. Our model both reproduces emergent behavior of human
drivers, such as lane change rate, while maintaining realistic control over
long time horizons.Comment: 8 pages, 6 figure
CoverNav: Cover Following Navigation Planning in Unstructured Outdoor Environment with Deep Reinforcement Learning
Autonomous navigation in offroad environments has been extensively studied in
the robotics field. However, navigation in covert situations where an
autonomous vehicle needs to remain hidden from outside observers remains an
underexplored area. In this paper, we propose a novel Deep Reinforcement
Learning (DRL) based algorithm, called CoverNav, for identifying covert and
navigable trajectories with minimal cost in offroad terrains and jungle
environments in the presence of observers. CoverNav focuses on unmanned ground
vehicles seeking shelters and taking covers while safely navigating to a
predefined destination. Our proposed DRL method computes a local cost map that
helps distinguish which path will grant the maximal covertness while
maintaining a low cost trajectory using an elevation map generated from 3D
point cloud data, the robot's pose, and directed goal information. CoverNav
helps robot agents to learn the low elevation terrain using a reward function
while penalizing it proportionately when it experiences high elevation. If an
observer is spotted, CoverNav enables the robot to select natural obstacles
(e.g., rocks, houses, disabled vehicles, trees, etc.) and use them as shelters
to hide behind. We evaluate CoverNav using the Unity simulation environment and
show that it guarantees dynamically feasible velocities in the terrain when fed
with an elevation map generated by another DRL based navigation algorithm.
Additionally, we evaluate CoverNav's effectiveness in achieving a maximum goal
distance of 12 meters and its success rate in different elevation scenarios
with and without cover objects. We observe competitive performance comparable
to state of the art (SOTA) methods without compromising accuracy
- …