40 research outputs found
Improving Automated Driving through Planning with Human Internal States
This work examines the hypothesis that partially observable Markov decision
process (POMDP) planning with human driver internal states can significantly
improve both safety and efficiency in autonomous freeway driving. We evaluate
this hypothesis in a simulated scenario where an autonomous car must safely
perform three lane changes in rapid succession. Approximate POMDP solutions are
obtained through the partially observable Monte Carlo planning with observation
widening (POMCPOW) algorithm. This approach outperforms over-confident and
conservative MDP baselines and matches or outperforms QMDP. Relative to the MDP
baselines, POMCPOW typically cuts the rate of unsafe situations in half or
increases the success rate by 50%.Comment: Preprint before submission to IEEE Transactions on Intelligent
Transportation Systems. arXiv admin note: text overlap with arXiv:1702.0085
Online algorithms for POMDPs with continuous state, action, and observation spaces
Online solvers for partially observable Markov decision processes have been
applied to problems with large discrete state spaces, but continuous state,
action, and observation spaces remain a challenge. This paper begins by
investigating double progressive widening (DPW) as a solution to this
challenge. However, we prove that this modification alone is not sufficient
because the belief representations in the search tree collapse to a single
particle causing the algorithm to converge to a policy that is suboptimal
regardless of the computation time. This paper proposes and evaluates two new
algorithms, POMCPOW and PFT-DPW, that overcome this deficiency by using
weighted particle filtering. Simulation results show that these modifications
allow the algorithms to be successful where previous approaches fail.Comment: Added Multilane sectio
Leveraging Counterfactual Paths for Contrastive Explanations of POMDP Policies
As humans come to rely on autonomous systems more, ensuring the transparency
of such systems is important to their continued adoption. Explainable
Artificial Intelligence (XAI) aims to reduce confusion and foster trust in
systems by providing explanations of agent behavior. Partially observable
Markov decision processes (POMDPs) provide a flexible framework capable of
reasoning over transition and state uncertainty, while also being amenable to
explanation. This work investigates the use of user-provided counterfactuals to
generate contrastive explanations of POMDP policies. Feature expectations are
used as a means of contrasting the performance of these policies. We
demonstrate our approach in a Search and Rescue (SAR) setting. We analyze and
discuss the associated challenges through two case studies.Comment: 5 pages, 1 figur
Planning with SiMBA: Motion Planning under Uncertainty for Temporal Goals using Simplified Belief Guides
This paper presents a new multi-layered algorithm for motion planning under
motion and sensing uncertainties for Linear Temporal Logic specifications. We
propose a technique to guide a sampling-based search tree in the combined task
and belief space using trajectories from a simplified model of the system, to
make the problem computationally tractable. Our method eliminates the need to
construct fine and accurate finite abstractions. We prove correctness and
probabilistic completeness of our algorithm, and illustrate the benefits of our
approach on several case studies. Our results show that guidance with a
simplified belief space model allows for significant speed-up in planning for
complex specifications.Comment: 8 pages, to appear in the IEEE International Conference on Robotics
and Automation (ICRA), 202
Human-Centered Autonomy for UAS Target Search
Current methods of deploying robots that operate in dynamic, uncertain
environments, such as Uncrewed Aerial Systems in search \& rescue missions,
require nearly continuous human supervision for vehicle guidance and operation.
These methods do not consider high-level mission context resulting in
cumbersome manual operation or inefficient exhaustive search patterns. We
present a human-centered autonomous framework that infers geospatial mission
context through dynamic feature sets, which then guides a probabilistic target
search planner. Operators provide a set of diverse inputs, including priority
definition, spatial semantic information about ad-hoc geographical areas, and
reference waypoints, which are probabilistically fused with geographical
database information and condensed into a geospatial distribution representing
an operator's preferences over an area. An online, POMDP-based planner,
optimized for target searching, is augmented with this reward map to generate
an operator-constrained policy. Our results, simulated based on input from five
professional rescuers, display effective task mental model alignment, 18\% more
victim finds, and 15 times more efficient guidance plans then current
operational methods.Comment: Extended version to ICRA conference submission. 9 pages, 5 figure
