7 research outputs found
Learning to Segment and Represent Motion Primitives from Driving Data for Motion Planning Applications
Developing an intelligent vehicle which can perform human-like actions
requires the ability to learn basic driving skills from a large amount of
naturalistic driving data. The algorithms will become efficient if we could
decompose the complex driving tasks into motion primitives which represent the
elementary compositions of driving skills. Therefore, the purpose of this paper
is to segment unlabeled trajectory data into a library of motion primitives. By
applying a probabilistic inference based on an iterative
Expectation-Maximization algorithm, our method segments the collected
trajectories while learning a set of motion primitives represented by the
dynamic movement primitives. The proposed method utilizes the mutual
dependencies between the segmentation and representation of motion primitives
and the driving-specific based initial segmentation. By utilizing this mutual
dependency and the initial condition, this paper presents how we can enhance
the performance of both the segmentation and the motion primitive library
establishment. We also evaluate the applicability of the primitive
representation method to imitation learning and motion planning algorithms. The
model is trained and validated by using the driving data collected from the
Beijing Institute of Technology intelligent vehicle platform. The results show
that the proposed approach can find the proper segmentation and establish the
motion primitive library simultaneously
A Tempt to Unify Heterogeneous Driving Databases using Traffic Primitives
A multitude of publicly-available driving datasets and data platforms have
been raised for autonomous vehicles (AV). However, the heterogeneities of
databases in size, structure and driving context make existing datasets
practically ineffective due to a lack of uniform frameworks and searchable
indexes. In order to overcome these limitations on existing public datasets,
this paper proposes a data unification framework based on traffic primitives
with ability to automatically unify and label heterogeneous traffic data. This
is achieved by two steps: 1) Carefully arrange raw multidimensional time series
driving data into a relational database and then 2) automatically extract
labeled and indexed traffic primitives from traffic data through a Bayesian
nonparametric learning method. Finally, we evaluate the effectiveness of our
developed framework using the collected real vehicle data.Comment: 6 pages, 7 figures, 1 table, ITSC 201
Symbol Emergence in Robotics: A Survey
Humans can learn the use of language through physical interaction with their
environment and semiotic communication with other people. It is very important
to obtain a computational understanding of how humans can form a symbol system
and obtain semiotic skills through their autonomous mental development.
Recently, many studies have been conducted on the construction of robotic
systems and machine-learning methods that can learn the use of language through
embodied multimodal interaction with their environment and other systems.
Understanding human social interactions and developing a robot that can
smoothly communicate with human users in the long term, requires an
understanding of the dynamics of symbol systems and is crucially important. The
embodied cognition and social interaction of participants gradually change a
symbol system in a constructive manner. In this paper, we introduce a field of
research called symbol emergence in robotics (SER). SER is a constructive
approach towards an emergent symbol system. The emergent symbol system is
socially self-organized through both semiotic communications and physical
interactions with autonomous cognitive developmental agents, i.e., humans and
developmental robots. Specifically, we describe some state-of-art research
topics concerning SER, e.g., multimodal categorization, word discovery, and a
double articulation analysis, that enable a robot to obtain words and their
embodied meanings from raw sensory--motor information, including visual
information, haptic information, auditory information, and acoustic speech
signals, in a totally unsupervised manner. Finally, we suggest future
directions of research in SER.Comment: submitted to Advanced Robotic
Modeling and Prediction of Driving Behaviors Using a Nonparametric Bayesian Method with AR Models
To develop a new generation advanced driver assistance system that avoids a dangerous condition in advance, we need to predict driving behaviors. Since a nonparametric Bayesian method with a two-level structure successfully predicted the symbolized behaviors only, we applied a nonparametric Bayesian method with linear dynamical systems to predicting the driving behavior. The method called the beta process autoregressive hidden Markov model (BP-AR-HMM) segments driving behaviors into states each of which corresponds to an AR model and it predicts future behaviors using the estimated future state sequence and the dynamical systems therein. Here, the segmentation as well as the parameters of the dynamical systems are determined using given training data in an unsupervised way. We carried out experiments with real driving data and found that the BP-AR-HMM predicted driving behaviors better than other methods
Spatiotemporal Learning of Multivehicle Interaction Patterns in Lane-Change Scenarios
Interpretation of common-yet-challenging interaction scenarios can benefit
well-founded decisions for autonomous vehicles. Previous research achieved this
using their prior knowledge of specific scenarios with predefined models,
limiting their adaptive capabilities. This paper describes a Bayesian
nonparametric approach that leverages continuous (i.e., Gaussian processes) and
discrete (i.e., Dirichlet processes) stochastic processes to reveal underlying
interaction patterns of the ego vehicle with other nearby vehicles. Our model
relaxes dependency on the number of surrounding vehicles by developing an
acceleration-sensitive velocity field based on Gaussian processes. The
experiment results demonstrate that the velocity field can represent the
spatial interactions between the ego vehicle and its surroundings. Then, a
discrete Bayesian nonparametric model, integrating Dirichlet processes and
hidden Markov models, is developed to learn the interaction patterns over the
temporal space by segmenting and clustering the sequential interaction data
into interpretable granular patterns automatically. We then evaluate our
approach in the highway lane-change scenarios using the highD dataset collected
from real-world settings. Results demonstrate that our proposed Bayesian
nonparametric approach provides an insight into the complicated lane-change
interactions of the ego vehicle with multiple surrounding traffic participants
based on the interpretable interaction patterns and their transition properties
in temporal relationships. Our proposed approach sheds light on efficiently
analyzing other kinds of multi-agent interactions, such as vehicle-pedestrian
interactions. View the demos via https://youtu.be/z_vf9UHtdAM.Comment: for the supplements, see
https://chengyuan-zhang.github.io/Multivehicle-Interaction