31,006 research outputs found
Search-based optimal motion planning for automated driving
This paper presents a framework for fast and robust motion planning designed
to facilitate automated driving. The framework allows for real-time computation
even for horizons of several hundred meters and thus enabling automated driving
in urban conditions. This is achieved through several features. Firstly, a
convenient geometrical representation of both the search space and driving
constraints enables the use of classical path planning approach. Thus, a wide
variety of constraints can be tackled simultaneously (other vehicles, traffic
lights, etc.). Secondly, an exact cost-to-go map, obtained by solving a relaxed
problem, is then used by A*-based algorithm with model predictive flavour in
order to compute the optimal motion trajectory. The algorithm takes into
account both distance and time horizons. The approach is validated within a
simulation study with realistic traffic scenarios. We demonstrate the
capability of the algorithm to devise plans both in fast and slow driving
conditions, even when full stop is required.Comment: Preprint accepted to 2018 IEEE/RSJ International Conference on
Intelligent Robots and Systems (IROS 2018). A supplementary video is
available at https://youtu.be/D5XJ5ncSuq
Driving with Style: Inverse Reinforcement Learning in General-Purpose Planning for Automated Driving
Behavior and motion planning play an important role in automated driving.
Traditionally, behavior planners instruct local motion planners with predefined
behaviors. Due to the high scene complexity in urban environments,
unpredictable situations may occur in which behavior planners fail to match
predefined behavior templates. Recently, general-purpose planners have been
introduced, combining behavior and local motion planning. These general-purpose
planners allow behavior-aware motion planning given a single reward function.
However, two challenges arise: First, this function has to map a complex
feature space into rewards. Second, the reward function has to be manually
tuned by an expert. Manually tuning this reward function becomes a tedious
task. In this paper, we propose an approach that relies on human driving
demonstrations to automatically tune reward functions. This study offers
important insights into the driving style optimization of general-purpose
planners with maximum entropy inverse reinforcement learning. We evaluate our
approach based on the expected value difference between learned and
demonstrated policies. Furthermore, we compare the similarity of human driven
trajectories with optimal policies of our planner under learned and
expert-tuned reward functions. Our experiments show that we are able to learn
reward functions exceeding the level of manual expert tuning without prior
domain knowledge.Comment: Appeared at IROS 2019. Accepted version. Added/updated footnote,
minor correction in preliminarie
Decentralized Cooperative Planning for Automated Vehicles with Hierarchical Monte Carlo Tree Search
Today's automated vehicles lack the ability to cooperate implicitly with
others. This work presents a Monte Carlo Tree Search (MCTS) based approach for
decentralized cooperative planning using macro-actions for automated vehicles
in heterogeneous environments. Based on cooperative modeling of other agents
and Decoupled-UCT (a variant of MCTS), the algorithm evaluates the
state-action-values of each agent in a cooperative and decentralized manner,
explicitly modeling the interdependence of actions between traffic
participants. Macro-actions allow for temporal extension over multiple time
steps and increase the effective search depth requiring fewer iterations to
plan over longer horizons. Without predefined policies for macro-actions, the
algorithm simultaneously learns policies over and within macro-actions. The
proposed method is evaluated under several conflict scenarios, showing that the
algorithm can achieve effective cooperative planning with learned macro-actions
in heterogeneous environments
Automating Vehicles by Deep Reinforcement Learning using Task Separation with Hill Climbing
Within the context of autonomous driving a model-based reinforcement learning
algorithm is proposed for the design of neural network-parameterized
controllers. Classical model-based control methods, which include sampling- and
lattice-based algorithms and model predictive control, suffer from the
trade-off between model complexity and computational burden required for the
online solution of expensive optimization or search problems at every short
sampling time. To circumvent this trade-off, a 2-step procedure is motivated:
first learning of a controller during offline training based on an arbitrarily
complicated mathematical system model, before online fast feedforward
evaluation of the trained controller. The contribution of this paper is the
proposition of a simple gradient-free and model-based algorithm for deep
reinforcement learning using task separation with hill climbing (TSHC). In
particular, (i) simultaneous training on separate deterministic tasks with the
purpose of encoding many motion primitives in a neural network, and (ii) the
employment of maximally sparse rewards in combination with virtual velocity
constraints (VVCs) in setpoint proximity are advocated.Comment: 10 pages, 6 figures, 1 tabl
Decentralized Cooperative Planning for Automated Vehicles with Continuous Monte Carlo Tree Search
Urban traffic scenarios often require a high degree of cooperation between
traffic participants to ensure safety and efficiency. Observing the behavior of
others, humans infer whether or not others are cooperating. This work aims to
extend the capabilities of automated vehicles, enabling them to cooperate
implicitly in heterogeneous environments. Continuous actions allow for
arbitrary trajectories and hence are applicable to a much wider class of
problems than existing cooperative approaches with discrete action spaces.
Based on cooperative modeling of other agents, Monte Carlo Tree Search (MCTS)
in conjunction with Decoupled-UCT evaluates the action-values of each agent in
a cooperative and decentralized way, respecting the interdependence of actions
among traffic participants. The extension to continuous action spaces is
addressed by incorporating novel MCTS-specific enhancements for efficient
search space exploration. The proposed algorithm is evaluated under different
scenarios, showing that the algorithm is able to achieve effective cooperative
planning and generate solutions egocentric planning fails to identify
- …