9,547 research outputs found
Towards Full Automated Drive in Urban Environments: A Demonstration in GoMentum Station, California
Each year, millions of motor vehicle traffic accidents all over the world
cause a large number of fatalities, injuries and significant material loss.
Automated Driving (AD) has potential to drastically reduce such accidents. In
this work, we focus on the technical challenges that arise from AD in urban
environments. We present the overall architecture of an AD system and describe
in detail the perception and planning modules. The AD system, built on a
modified Acura RLX, was demonstrated in a course in GoMentum Station in
California. We demonstrated autonomous handling of 4 scenarios: traffic lights,
cross-traffic at intersections, construction zones and pedestrians. The AD
vehicle displayed safe behavior and performed consistently in repeated
demonstrations with slight variations in conditions. Overall, we completed 44
runs, encompassing 110km of automated driving with only 3 cases where the
driver intervened the control of the vehicle, mostly due to error in GPS
positioning. Our demonstration showed that robust and consistent behavior in
urban scenarios is possible, yet more investigation is necessary for full scale
roll-out on public roads.Comment: Accepted to Intelligent Vehicles Conference (IV 2017
Implementation and Evaluation of a Cooperative Vehicle-to-Pedestrian Safety Application
While the development of Vehicle-to-Vehicle (V2V) safety applications based
on Dedicated Short-Range Communications (DSRC) has been extensively undergoing
standardization for more than a decade, such applications are extremely missing
for Vulnerable Road Users (VRUs). Nonexistence of collaborative systems between
VRUs and vehicles was the main reason for this lack of attention. Recent
developments in Wi-Fi Direct and DSRC-enabled smartphones are changing this
perspective. Leveraging the existing V2V platforms, we propose a new framework
using a DSRC-enabled smartphone to extend safety benefits to VRUs. The
interoperability of applications between vehicles and portable DSRC enabled
devices is achieved through the SAE J2735 Personal Safety Message (PSM).
However, considering the fact that VRU movement dynamics, response times, and
crash scenarios are fundamentally different from vehicles, a specific framework
should be designed for VRU safety applications to study their performance. In
this article, we first propose an end-to-end Vehicle-to-Pedestrian (V2P)
framework to provide situational awareness and hazard detection based on the
most common and injury-prone crash scenarios. The details of our VRU safety
module, including target classification and collision detection algorithms, are
explained next. Furthermore, we propose and evaluate a mitigating solution for
congestion and power consumption issues in such systems. Finally, the whole
system is implemented and analyzed for realistic crash scenarios
Limited Visibility and Uncertainty Aware Motion Planning for Automated Driving
Adverse weather conditions and occlusions in urban environments result in
impaired perception. The uncertainties are handled in different modules of an
automated vehicle, ranging from sensor level over situation prediction until
motion planning. This paper focuses on motion planning given an uncertain
environment model with occlusions. We present a method to remain collision free
for the worst-case evolution of the given scene. We define criteria that
measure the available margins to a collision while considering visibility and
interactions, and consequently integrate conditions that apply these criteria
into an optimization-based motion planner. We show the generality of our method
by validating it in several distinct urban scenarios
Egocentric Vision-based Future Vehicle Localization for Intelligent Driving Assistance Systems
Predicting the future location of vehicles is essential for safety-critical
applications such as advanced driver assistance systems (ADAS) and autonomous
driving. This paper introduces a novel approach to simultaneously predict both
the location and scale of target vehicles in the first-person (egocentric) view
of an ego-vehicle. We present a multi-stream recurrent neural network (RNN)
encoder-decoder model that separately captures both object location and scale
and pixel-level observations for future vehicle localization. We show that
incorporating dense optical flow improves prediction results significantly
since it captures information about motion as well as appearance change. We
also find that explicitly modeling future motion of the ego-vehicle improves
the prediction accuracy, which could be especially beneficial in intelligent
and automated vehicles that have motion planning capability. To evaluate the
performance of our approach, we present a new dataset of first-person videos
collected from a variety of scenarios at road intersections, which are
particularly challenging moments for prediction because vehicle trajectories
are diverse and dynamic.Comment: To appear on ICRA 201
RUR53: an Unmanned Ground Vehicle for Navigation, Recognition and Manipulation
This paper proposes RUR53: an Unmanned Ground Vehicle able to autonomously
navigate through, identify, and reach areas of interest; and there recognize,
localize, and manipulate work tools to perform complex manipulation tasks. The
proposed contribution includes a modular software architecture where each
module solves specific sub-tasks and that can be easily enlarged to satisfy new
requirements. Included indoor and outdoor tests demonstrate the capability of
the proposed system to autonomously detect a target object (a panel) and
precisely dock in front of it while avoiding obstacles. They show it can
autonomously recognize and manipulate target work tools (i.e., wrenches and
valve stems) to accomplish complex tasks (i.e., use a wrench to rotate a valve
stem). A specific case study is described where the proposed modular
architecture lets easy switch to a semi-teleoperated mode. The paper
exhaustively describes description of both the hardware and software setup of
RUR53, its performance when tests at the 2017 Mohamed Bin Zayed International
Robotics Challenge, and the lessons we learned when participating at this
competition, where we ranked third in the Gran Challenge in collaboration with
the Czech Technical University in Prague, the University of Pennsylvania, and
the University of Lincoln (UK).Comment: This article has been accepted for publication in Advanced Robotics,
published by Taylor & Franci
- …