4,638 research outputs found
Learning to Segment and Represent Motion Primitives from Driving Data for Motion Planning Applications
Developing an intelligent vehicle which can perform human-like actions
requires the ability to learn basic driving skills from a large amount of
naturalistic driving data. The algorithms will become efficient if we could
decompose the complex driving tasks into motion primitives which represent the
elementary compositions of driving skills. Therefore, the purpose of this paper
is to segment unlabeled trajectory data into a library of motion primitives. By
applying a probabilistic inference based on an iterative
Expectation-Maximization algorithm, our method segments the collected
trajectories while learning a set of motion primitives represented by the
dynamic movement primitives. The proposed method utilizes the mutual
dependencies between the segmentation and representation of motion primitives
and the driving-specific based initial segmentation. By utilizing this mutual
dependency and the initial condition, this paper presents how we can enhance
the performance of both the segmentation and the motion primitive library
establishment. We also evaluate the applicability of the primitive
representation method to imitation learning and motion planning algorithms. The
model is trained and validated by using the driving data collected from the
Beijing Institute of Technology intelligent vehicle platform. The results show
that the proposed approach can find the proper segmentation and establish the
motion primitive library simultaneously
End-to-end Driving via Conditional Imitation Learning
Deep networks trained on demonstrations of human driving have learned to
follow roads and avoid obstacles. However, driving policies trained via
imitation learning cannot be controlled at test time. A vehicle trained
end-to-end to imitate an expert cannot be guided to take a specific turn at an
upcoming intersection. This limits the utility of such systems. We propose to
condition imitation learning on high-level command input. At test time, the
learned driving policy functions as a chauffeur that handles sensorimotor
coordination but continues to respond to navigational commands. We evaluate
different architectures for conditional imitation learning in vision-based
driving. We conduct experiments in realistic three-dimensional simulations of
urban driving and on a 1/5 scale robotic truck that is trained to drive in a
residential area. Both systems drive based on visual input yet remain
responsive to high-level navigational commands. The supplementary video can be
viewed at https://youtu.be/cFtnflNe5fMComment: Published at the International Conference on Robotics and Automation
(ICRA), 201
Motion Planning for Autonomous Driving: The State of the Art and Future Perspectives
Thanks to the augmented convenience, safety advantages, and potential
commercial value, Intelligent vehicles (IVs) have attracted wide attention
throughout the world. Although a few autonomous driving unicorns assert that
IVs will be commercially deployable by 2025, their implementation is still
restricted to small-scale validation due to various issues, among which precise
computation of control commands or trajectories by planning methods remains a
prerequisite for IVs. This paper aims to review state-of-the-art planning
methods, including pipeline planning and end-to-end planning methods. In terms
of pipeline methods, a survey of selecting algorithms is provided along with a
discussion of the expansion and optimization mechanisms, whereas in end-to-end
methods, the training approaches and verification scenarios of driving tasks
are points of concern. Experimental platforms are reviewed to facilitate
readers in selecting suitable training and validation methods. Finally, the
current challenges and future directions are discussed. The side-by-side
comparison presented in this survey not only helps to gain insights into the
strengths and limitations of the reviewed methods but also assists with
system-level design choices.Comment: 20 pages, 14 figures and 5 table
- …