13 research outputs found
Simulation-based reinforcement learning for real-world autonomous driving
We use reinforcement learning in simulation to obtain a driving system
controlling a full-size real-world vehicle. The driving policy takes RGB images
from a single camera and their semantic segmentation as input. We use mostly
synthetic data, with labelled real-world data appearing only in the training of
the segmentation network.
Using reinforcement learning in simulation and synthetic data is motivated by
lowering costs and engineering effort.
In real-world experiments we confirm that we achieved successful sim-to-real
policy transfer. Based on the extensive evaluation, we analyze how design
decisions about perception, control, and training impact the real-world
performance
Comparative study of optimization algorithms on convolutional network for autonomous driving
he last 10 years have been the decade of autonomous vehicles. Advances in intelligent sensors and control schemes have shown the possibility of real applications. Deep learning, and in particular convolutional networks have become a fundamental tool in the solution of problems related to environment identification, path planning, vehicle behavior, and motion control. In this paper, we perform a comparative study of the most used optimization strategies on the convolutional architecture residual neural network (ResNet) for an autonomous driving problem as a previous step to the development of an intelligent sensor. This sensor, part of our research in reactive systems for autonomous vehicles, aims to become a system for direct mapping of sensory information to control actions from real-time images of the environment. The optimization techniques analyzed include stochastic gradient descent (SGD), adaptive gradient (Adagrad), adaptive learning rate (Adadelta), root mean square propagation (RMSProp), Adamax, adaptive moment estimation (Adam), nesterov-accelerated adaptive moment estimation (Nadam), and follow the regularized leader (Ftrl). The training of the deep model is evaluated in terms of convergence, accuracy, recall, and F1-score metrics. Preliminary results show a better performance of the deep network when using the SGD function as an optimizer, while the Ftrl function presents the poorest performances
Waymax: An Accelerated, Data-Driven Simulator for Large-Scale Autonomous Driving Research
Simulation is an essential tool to develop and benchmark autonomous vehicle
planning software in a safe and cost-effective manner. However, realistic
simulation requires accurate modeling of nuanced and complex multi-agent
interactive behaviors. To address these challenges, we introduce Waymax, a new
data-driven simulator for autonomous driving in multi-agent scenes, designed
for large-scale simulation and testing. Waymax uses publicly-released,
real-world driving data (e.g., the Waymo Open Motion Dataset) to initialize or
play back a diverse set of multi-agent simulated scenarios. It runs entirely on
hardware accelerators such as TPUs/GPUs and supports in-graph simulation for
training, making it suitable for modern large-scale, distributed machine
learning workflows. To support online training and evaluation, Waymax includes
several learned and hard-coded behavior models that allow for realistic
interaction within simulation. To supplement Waymax, we benchmark a suite of
popular imitation and reinforcement learning algorithms with ablation studies
on different design decisions, where we highlight the effectiveness of routes
as guidance for planning agents and the ability of RL to overfit against
simulated agents
LEyes: A Lightweight Framework for Deep Learning-Based Eye Tracking using Synthetic Eye Images
Deep learning has bolstered gaze estimation techniques, but real-world
deployment has been impeded by inadequate training datasets. This problem is
exacerbated by both hardware-induced variations in eye images and inherent
biological differences across the recorded participants, leading to both
feature and pixel-level variance that hinders the generalizability of models
trained on specific datasets. While synthetic datasets can be a solution, their
creation is both time and resource-intensive. To address this problem, we
present a framework called Light Eyes or "LEyes" which, unlike conventional
photorealistic methods, only models key image features required for video-based
eye tracking using simple light distributions. LEyes facilitates easy
configuration for training neural networks across diverse gaze-estimation
tasks. We demonstrate that models trained using LEyes are consistently on-par
or outperform other state-of-the-art algorithms in terms of pupil and CR
localization across well-known datasets. In addition, a LEyes trained model
outperforms the industry standard eye tracker using significantly more
cost-effective hardware. Going forward, we are confident that LEyes will
revolutionize synthetic data generation for gaze estimation models, and lead to
significant improvements of the next generation video-based eye trackers.Comment: 32 pages, 8 figure