594 research outputs found
Did You Miss the Sign? A False Negative Alarm System for Traffic Sign Detectors
Object detection is an integral part of an autonomous vehicle for its
safety-critical and navigational purposes. Traffic signs as objects play a
vital role in guiding such systems. However, if the vehicle fails to locate any
critical sign, it might make a catastrophic failure. In this paper, we propose
an approach to identify traffic signs that have been mistakenly discarded by
the object detector. The proposed method raises an alarm when it discovers a
failure by the object detector to detect a traffic sign. This approach can be
useful to evaluate the performance of the detector during the deployment phase.
We trained a single shot multi-box object detector to detect traffic signs and
used its internal features to train a separate false negative detector (FND).
During deployment, FND decides whether the traffic sign detector (TSD) has
missed a sign or not. We are using precision and recall to measure the accuracy
of FND in two different datasets. For 80% recall, FND has achieved 89.9%
precision in Belgium Traffic Sign Detection dataset and 90.8% precision in
German Traffic Sign Recognition Benchmark dataset respectively. To the best of
our knowledge, our method is the first to tackle this critical aspect of false
negative detection in robotic vision. Such a fail-safe mechanism for object
detection can improve the engagement of robotic vision systems in our daily
life.Comment: Submitted to the 2019 IEEE/RSJ International Conference on
Intelligent Robots and Systems (IROS 2019
Dropout Distillation for Efficiently Estimating Model Confidence
We propose an efficient way to output better calibrated uncertainty scores
from neural networks. The Distilled Dropout Network (DDN) makes standard
(non-Bayesian) neural networks more introspective by adding a new training loss
which prevents them from being overconfident. Our method is more efficient than
Bayesian neural networks or model ensembles which, despite providing more
reliable uncertainty scores, are more cumbersome to train and slower to test.
We evaluate DDN on the the task of image classification on the CIFAR-10 dataset
and show that our calibration results are competitive even when compared to 100
Monte Carlo samples from a dropout network while they also increase the
classification accuracy. We also propose better calibration within the state of
the art Faster R-CNN object detection framework and show, using the COCO
dataset, that DDN helps train better calibrated object detectors
Recommended from our members
Explainable and Advisable Learning for Self-driving Vehicles
Deep neural perception and control networks are likely to be a key component of self-driving vehicles. These models need to be explainable - they should provide easy-to-interpret rationales for their behavior - so that passengers, insurance companies, law enforcement, developers, etc., can understand what triggered a particular behavior. Explanations may be triggered by the neural controller, namely introspective explanations, or informed by the neural controller's output, namely rationalizations. Our work has focused on the challenge of generating introspective explanations of deep models for self-driving vehicles. In Chapter 3, we begin by exploring the use of visual explanations. These explanations take the form of real-time highlighted regions of an image that causally influence the network's output (steering control). In the first stage, we use a visual attention model to train a convolution network end-to-end from images to steering angle. The attention model highlights image regions that potentially influence the network's output. Some of these are true influences, but some are spurious. We then apply a causal filtering step to determine which input regions actually influence the output. This produces more succinct visual explanations and more accurately exposes the network's behavior. In Chapter 4, we add an attention-based video-to-text model to produce textual explanations of model actions, e.g. "the car slows down because the road is wet". The attention maps of controller and explanation model are aligned so that explanations are grounded in the parts of the scene that mattered to the controller. We explore two approaches to attention alignment, strong- and weak-alignment. These explainable systems represent an externalization of tacit knowledge. The network's opaque reasoning is simplified to a situation-specific dependence on a visible object in the image. This makes them brittle and potentially unsafe in situations that do not match training data. In Chapter 5, we propose to address this issue by augmenting training data with natural language advice from a human. Advice includes guidance about what to do and where to attend. We present the first step toward advice-giving, where we train an end-to-end vehicle controller that accepts advice. The controller adapts the way it attends to the scene (visual attention) and the control (steering and speed). Further, in Chapter 6, we propose a new approach that learns vehicle control with the help of long-term (global) human advice. Specifically, our system learns to summarize its visual observations in natural language, predict an appropriate action response (e.g. "I see a pedestrian crossing, so I stop"), and predict the controls, accordingly
Per-frame mAP Prediction for Continuous Performance Monitoring of Object Detection During Deployment
Performance monitoring of object detection is crucial for safety-critical
applications such as autonomous vehicles that operate under varying and complex
environmental conditions. Currently, object detectors are evaluated using
summary metrics based on a single dataset that is assumed to be representative
of all future deployment conditions. In practice, this assumption does not
hold, and the performance fluctuates as a function of the deployment
conditions. To address this issue, we propose an introspection approach to
performance monitoring during deployment without the need for ground truth
data. We do so by predicting when the per-frame mean average precision drops
below a critical threshold using the detector's internal features. We
quantitatively evaluate and demonstrate our method's ability to reduce risk by
trading off making an incorrect decision by raising the alarm and absenting
from detection
Introspective Perception for Mobile Robots
Perception algorithms that provide estimates of their uncertainty are crucial
to the development of autonomous robots that can operate in challenging and
uncontrolled environments. Such perception algorithms provide the means for
having risk-aware robots that reason about the probability of successfully
completing a task when planning. There exist perception algorithms that come
with models of their uncertainty; however, these models are often developed
with assumptions, such as perfect data associations, that do not hold in the
real world. Hence the resultant estimated uncertainty is a weak lower bound. To
tackle this problem we present introspective perception - a novel approach for
predicting accurate estimates of the uncertainty of perception algorithms
deployed on mobile robots. By exploiting sensing redundancy and consistency
constraints naturally present in the data collected by a mobile robot,
introspective perception learns an empirical model of the error distribution of
perception algorithms in the deployment environment and in an autonomously
supervised manner. In this paper, we present the general theory of
introspective perception and demonstrate successful implementations for two
different perception tasks. We provide empirical results on challenging
real-robot data for introspective stereo depth estimation and introspective
visual simultaneous localization and mapping and show that they learn to
predict their uncertainty with high accuracy and leverage this information to
significantly reduce state estimation errors for an autonomous mobile robot
Lidar-level localization with radar? The CFEAR approach to accurate, fast and robust large-scale radar odometry in diverse environments
This paper presents an accurate, highly efficient, and learning-free method
for large-scale odometry estimation using spinning radar, empirically found to
generalize well across very diverse environments -- outdoors, from urban to
woodland, and indoors in warehouses and mines - without changing parameters.
Our method integrates motion compensation within a sweep with one-to-many scan
registration that minimizes distances between nearby oriented surface points
and mitigates outliers with a robust loss function. Extending our previous
approach CFEAR, we present an in-depth investigation on a wider range of data
sets, quantifying the importance of filtering, resolution, registration cost
and loss functions, keyframe history, and motion compensation. We present a new
solving strategy and configuration that overcomes previous issues with sparsity
and bias, and improves our state-of-the-art by 38%, thus, surprisingly,
outperforming radar SLAM and approaching lidar SLAM. The most accurate
configuration achieves 1.09% error at 5Hz on the Oxford benchmark, and the
fastest achieves 1.79% error at 160Hz.Comment: Accepted for publication in Transactions on Robotics. Edited
2022-11-07: Updated affiliation and citatio
- …