1,279 research outputs found
Monocular 3D Object Detection via Ego View-to-Bird’s Eye View Translation
The advanced development in autonomous agents like self-driving cars can be attributed to computer vision, a branch of artificial intelligence that enables software to understand the content of image and video. These autonomous agents require a three-dimensional modelling of its surrounding in order to operate reliably in the real-world. Despite the significant progress of 2D object detectors, they have a critical limitation in location sensitive applications as they do not provide accurate physical information of objects in 3D space. 3D object detection is a promising topic that can provide relevant solutions which could improve existing 2D based applications. Due to the advancements in deep learning methods and relevant datasets, the task of 3D scene understanding has evolved greatly in the past few years. 3D object detection and localization are crucial in autonomous driving tasks such as obstacle avoidance, path planning and motion control. Traditionally, there have been successful methods towards 3D object detection but they rely on highly expensive 3D LiDAR sensors for accurate depth information. On the other hand, 3D object detection from single monocular images is inexpensive but lacks in accuracy. The primary reason for such a disparity in performance is that the monocular image-based methods attempt at inferring 3D information from 2D images. In this work, we try to bridge the performance gap observed in single image input by introducing different mapping strategies between the 2D image data and its corresponding 3D representation and use it to perform object detection in 3D. The performance of the proposed method is evaluated on the popular KITTI 3D object detection benchmark dataset
Active Perception using Light Curtains for Autonomous Driving
Most real-world 3D sensors such as LiDARs perform fixed scans of the entire
environment, while being decoupled from the recognition system that processes
the sensor data. In this work, we propose a method for 3D object recognition
using light curtains, a resource-efficient controllable sensor that measures
depth at user-specified locations in the environment. Crucially, we propose
using prediction uncertainty of a deep learning based 3D point cloud detector
to guide active perception. Given a neural network's uncertainty, we derive an
optimization objective to place light curtains using the principle of
maximizing information gain. Then, we develop a novel and efficient
optimization algorithm to maximize this objective by encoding the physical
constraints of the device into a constraint graph and optimizing with dynamic
programming. We show how a 3D detector can be trained to detect objects in a
scene by sequentially placing uncertainty-guided light curtains to successively
improve detection accuracy. Code and details can be found on the project
webpage: http://siddancha.github.io/projects/active-perception-light-curtains.Comment: Published at the European Conference on Computer Vision (ECCV), 202
Road Surface Feature Extraction and Reconstruction of Laser Point Clouds for Urban Environment
Automakers are developing end-to-end three-dimensional (3D) mapping system for Advanced Driver Assistance Systems (ADAS) and autonomous vehicles (AVs). Using geomatics, artificial intelligence, and SLAM (Simultaneous Localization and Mapping) systems to handle all stages of map creation, sensor calibration and alignment. It is crucial to have a system highly accurate and efficient as it is an essential part of vehicle controls. Such mapping requires significant resources to acquire geographic information (GIS and GPS), optical laser and radar spectroscopy, Lidar, and 3D modeling applications in order to extract roadway features (e.g., lane markings, traffic signs, road-edges) detailed enough to construct a “base map”. To keep this map current, it is necessary to update changes due to occurring events such as construction changes, traffic patterns, or growth of vegetation. The information of the road play a very important factor in road traffic safety and it is essential for for guiding autonomous vehicles (AVs), and prediction of upcoming road situations within AVs. The data size of the map is extensive due to the level of information provided with different sensor modalities for that reason a data optimization and extraction from three-dimensional (3D) mobile laser scanning (MLS) point clouds is presented in this thesis. The research shows the proposed hybrid filter configuration together with the dynamic developed mechanism provides significant reduction of the point cloud data with reduced computational or size constraints. The results obtained in this work are proven by a real-world system
- …