4,993 research outputs found
Anomaly Detection in Autonomous Driving: A Survey
Nowadays, there are outstanding strides towards a future with autonomous
vehicles on our roads. While the perception of autonomous vehicles performs
well under closed-set conditions, they still struggle to handle the unexpected.
This survey provides an extensive overview of anomaly detection techniques
based on camera, lidar, radar, multimodal and abstract object level data. We
provide a systematization including detection approach, corner case level,
ability for an online application, and further attributes. We outline the
state-of-the-art and point out current research gaps.Comment: Daniel Bogdoll and Maximilian Nitsche contributed equally. Accepted
for publication at CVPR 2022 WAD worksho
Deep reinforcement learning for drone navigation using sensor data
Mobile robots such as unmanned aerial vehicles (drones) can be used for surveillance, monitoring and data collection in buildings, infrastructure and environments. The importance of accurate and multifaceted monitoring is well known to identify problems early and prevent them escalating. This motivates the need for flexible, autonomous and powerful decision-making mobile robots. These systems need to be able to learn through fusing data from multiple sources. Until very recently, they have been task specific. In this paper, we describe a generic navigation algorithm that uses data from sensors on-board the drone to guide the drone to the site of the problem. In hazardous and safety-critical situations, locating problems accurately and rapidly is vital. We use the proximal policy optimisation deep reinforcement learning algorithm coupled with incremental curriculum learning and long short-term memory neural networks to implement our generic and adaptable navigation algorithm. We evaluate different configurations against a heuristic technique to demonstrate its accuracy and efficiency. Finally, we consider how safety of the drone could be assured by assessing how safely the drone would perform using our navigation algorithm in real-world scenarios
TractorEYE: Vision-based Real-time Detection for Autonomous Vehicles in Agriculture
Agricultural vehicles such as tractors and harvesters have for decades been able to navigate automatically and more efficiently using commercially available products such as auto-steering and tractor-guidance systems. However, a human operator is still required inside the vehicle to ensure the safety of vehicle and especially surroundings such as humans and animals. To get fully autonomous vehicles certified for farming, computer vision algorithms and sensor technologies must detect obstacles with equivalent or better than human-level performance. Furthermore, detections must run in real-time to allow vehicles to actuate and avoid collision.This thesis proposes a detection system (TractorEYE), a dataset (FieldSAFE), and procedures to fuse information from multiple sensor technologies to improve detection of obstacles and to generate a map. TractorEYE is a multi-sensor detection system for autonomous vehicles in agriculture. The multi-sensor system consists of three hardware synchronized and registered sensors (stereo camera, thermal camera and multi-beam lidar) mounted on/in a ruggedized and water-resistant casing. Algorithms have been developed to run a total of six detection algorithms (four for rgb camera, one for thermal camera and one for a Multi-beam lidar) and fuse detection information in a common format using either 3D positions or Inverse Sensor Models. A GPU powered computational platform is able to run detection algorithms online. For the rgb camera, a deep learning algorithm is proposed DeepAnomaly to perform real-time anomaly detection of distant, heavy occluded and unknown obstacles in agriculture. DeepAnomaly is -- compared to a state-of-the-art object detector Faster R-CNN -- for an agricultural use-case able to detect humans better and at longer ranges (45-90m) using a smaller memory footprint and 7.3-times faster processing. Low memory footprint and fast processing makes DeepAnomaly suitable for real-time applications running on an embedded GPU. FieldSAFE is a multi-modal dataset for detection of static and moving obstacles in agriculture. The dataset includes synchronized recordings from a rgb camera, stereo camera, thermal camera, 360-degree camera, lidar and radar. Precise localization and pose is provided using IMU and GPS. Ground truth of static and moving obstacles (humans, mannequin dolls, barrels, buildings, vehicles, and vegetation) are available as an annotated orthophoto and GPS coordinates for moving obstacles. Detection information from multiple detection algorithms and sensors are fused into a map using Inverse Sensor Models and occupancy grid maps. This thesis presented many scientific contribution and state-of-the-art within perception for autonomous tractors; this includes a dataset, sensor platform, detection algorithms and procedures to perform multi-sensor fusion. Furthermore, important engineering contributions to autonomous farming vehicles are presented such as easily applicable, open-source software packages and algorithms that have been demonstrated in an end-to-end real-time detection system. The contributions of this thesis have demonstrated, addressed and solved critical issues to utilize camera-based perception systems that are essential to make autonomous vehicles in agriculture a reality
GrASPE: Graph based Multimodal Fusion for Robot Navigation in Unstructured Outdoor Environments
We present a novel trajectory traversability estimation and planning
algorithm for robot navigation in complex outdoor environments. We incorporate
multimodal sensory inputs from an RGB camera, 3D LiDAR, and robot's odometry
sensor to train a prediction model to estimate candidate trajectories' success
probabilities based on partially reliable multi-modal sensor observations. We
encode high-dimensional multi-modal sensory inputs to low-dimensional feature
vectors using encoder networks and represent them as a connected graph to train
an attention-based Graph Neural Network (GNN) model to predict trajectory
success probabilities. We further analyze the image and point cloud data
separately to quantify sensor reliability to augment the weights of the feature
graph representation used in our GNN. During runtime, our model utilizes
multi-sensor inputs to predict the success probabilities of the trajectories
generated by a local planner to avoid potential collisions and failures. Our
algorithm demonstrates robust predictions when one or more sensor modalities
are unreliable or unavailable in complex outdoor environments. We evaluate our
algorithm's navigation performance using a Spot robot in real-world outdoor
environments
Local and Global Information in Obstacle Detection on Railway Tracks
Reliable obstacle detection on railways could help prevent collisions that
result in injuries and potentially damage or derail the train. Unfortunately,
generic object detectors do not have enough classes to account for all possible
scenarios, and datasets featuring objects on railways are challenging to
obtain. We propose utilizing a shallow network to learn railway segmentation
from normal railway images. The limited receptive field of the network prevents
overconfident predictions and allows the network to focus on the locally very
distinct and repetitive patterns of the railway environment. Additionally, we
explore the controlled inclusion of global information by learning to
hallucinate obstacle-free images. We evaluate our method on a custom dataset
featuring railway images with artificially augmented obstacles. Our proposed
method outperforms other learning-based baseline methods
- …