50 research outputs found
Game Theoretic Analysis of Road User Safety Scenarios Involving Autonomous Vehicles
Interactions between pedestrians, bikers, and human-driven vehicles have been
a major concern in traffic safety over the years. The upcoming age of
autonomous vehicles will further raise major problems on whether self-driving
cars can accurately avoid accidents; on the other hand, usability issues arise
on whether human-driven cars and pedestrians can dominate the road at the
expense of the autonomous vehicles which will be programmed to avoid accidents.
This paper proposes some game theoretical models applied to related traffic
scenarios. In the first two games the reciprocal influence between a pedestrian
and a vehicle (either autonomous or not) is analyzed, while the third game
investigates the intersection of two vehicles, possibly autonomous. The games
have been simulated in order to demonstrate the theoretical analysis and the
predicted behaviors. These investigations can shed new lights on how novel
urban traffic regulations could be required to allow for a better interaction
of vehicles and a general improved management of traffic and communication
vehicular networks.Comment: Accepted at 'IEEE International Symposium on Personal, Indoor and
Mobile Radio Communications' 9-12 September 2018 - Bologna, Italy. Special
Session on 'Wireless Technologies for Connected and Autonomous Vehicles'. 7
pages, 5 figure
A Model for Every User and Budget: Label-Free and Personalized Mixed-Precision Quantization
Recent advancement in Automatic Speech Recognition (ASR) has produced large
AI models, which become impractical for deployment in mobile devices. Model
quantization is effective to produce compressed general-purpose models, however
such models may only be deployed to a restricted sub-domain of interest. We
show that ASR models can be personalized during quantization while relying on
just a small set of unlabelled samples from the target domain. To this end, we
propose myQASR, a mixed-precision quantization method that generates tailored
quantization schemes for diverse users under any memory requirement with no
fine-tuning. myQASR automatically evaluates the quantization sensitivity of
network layers by analysing the full-precision activation values. We are then
able to generate a personalised mixed-precision quantization scheme for any
pre-determined memory budget. Results for large-scale ASR models show how
myQASR improves performance for specific genders, languages, and speakers.Comment: INTERSPEECH 202
Online Continual Learning in Keyword Spotting for Low-Resource Devices via Pooling High-Order Temporal Statistics
Keyword Spotting (KWS) models on embedded devices should adapt fast to new
user-defined words without forgetting previous ones. Embedded devices have
limited storage and computational resources, thus, they cannot save samples or
update large models. We consider the setup of embedded online continual
learning (EOCL), where KWS models with frozen backbone are trained to
incrementally recognize new words from a non-repeated stream of samples, seen
one at a time. To this end, we propose Temporal Aware Pooling (TAP) which
constructs an enriched feature space computing high-order moments of speech
features extracted by a pre-trained backbone. Our method, TAP-SLDA, updates a
Gaussian model for each class on the enriched feature space to effectively use
audio representations. In experimental analyses, TAP-SLDA outperforms
competitors on several setups, backbones, and baselines, bringing a relative
average gain of 11.3% on the GSC dataset.Comment: INTERSPEECH 202
Latent Space Regularization for Unsupervised Domain Adaptation in Semantic Segmentation
Deep convolutional neural networks for semantic segmentation achieve
outstanding accuracy, however they also have a couple of major drawbacks:
first, they do not generalize well to distributions slightly different from the
one of the training data; second, they require a huge amount of labeled data
for their optimization. In this paper, we introduce feature-level space-shaping
regularization strategies to reduce the domain discrepancy in semantic
segmentation. In particular, for this purpose we jointly enforce a clustering
objective, a perpendicularity constraint and a norm alignment goal on the
feature vectors corresponding to source and target samples. Additionally, we
propose a novel measure able to capture the relative efficacy of an adaptation
strategy compared to supervised training. We verify the effectiveness of such
methods in the autonomous driving setting achieving state-of-the-art results in
multiple synthetic-to-real road scenes benchmarks.Comment: Accepted at CVPR-WAD 2021, 11 pages, 7 figures, 1 table
Adversarial Learning and Self-Teaching Techniques for Domain Adaptation in Semantic Segmentation
Deep learning techniques have been widely used in autonomous driving systems
for the semantic understanding of urban scenes. However, they need a huge
amount of labeled data for training, which is difficult and expensive to
acquire. A recently proposed workaround is to train deep networks using
synthetic data, but the domain shift between real world and synthetic
representations limits the performance. In this work, a novel Unsupervised
Domain Adaptation (UDA) strategy is introduced to solve this issue. The
proposed learning strategy is driven by three components: a standard supervised
learning loss on labeled synthetic data; an adversarial learning module that
exploits both labeled synthetic data and unlabeled real data; finally, a
self-teaching strategy applied to unlabeled data. The last component exploits a
region growing framework guided by the segmentation confidence. Furthermore, we
weighted this component on the basis of the class frequencies to enhance the
performance on less common classes. Experimental results prove the
effectiveness of the proposed strategy in adapting a segmentation network
trained on synthetic datasets, like GTA5 and SYNTHIA, to real world datasets
like Cityscapes and Mapillary.Comment: Accepted at IEEE Transactions on Intelligent Vehicles (T-IV) 10
pages, 2 figures, 7 table