42,558 research outputs found
Object Detection from a Vehicle Using Deep Learning Network and Future Integration with Multi-Sensor Fusion Algorithm
Accuracy in detecting a moving object is critical to autonomous driving or advanced driver assistance systems (ADAS). By including the object classification from multiple sensor detections, the model of the object or environment can be identified more accurately. The critical parameters involved in improving the accuracy are the size and the speed of the moving object. All sensor data are to be used in defining a composite object representation so that it could be used for the class information in the core object’s description. This composite data can then be used by a deep learning network for complete perception fusion in order to solve the detection and tracking of moving objects problem. Camera image data from subsequent frames along the time axis in conjunction with the speed and size of the object will further contribute in developing better recognition algorithms. In this paper, we present preliminary results using only camera images for detecting various objects using deep learning network, as a first step toward multi-sensor fusion algorithm development. The simulation experiments based on camera images show encouraging results where the proposed deep learning network based detection algorithm was able to detect various objects with certain degree of confidence. A laboratory experimental setup is being commissioned where three different types of sensors, a digital camera with 8 megapixel resolution, a LIDAR with 40m range, and ultrasonic distance transducer sensors will be used for multi-sensor fusion to identify the object in real-time
Fusion of Noisy Multi-sensor Imagery
Interest in fusing multiple sensor data for both military and civil applications has beengrowing. Some of the important applications integrate image information from multiple sensorsto aid in navigation guidance, object detection and recognition, medical diagnosis, datacompression, etc. While, human beings may visually inspect various images and integrateinformation, it is of interest to develop algorithms that can fuse various input imagery to producea composite image. Fusion of images from various sensor modalities is expected to produce anoutput that captures all the relevant information in the input. The standard multi-resolution-based edge fusion scheme has been reviewed in this paper. A theoretical framework is given forthis edge fusion method by showing how edge fusion can be framed as information maximisation.However, the presence of noise complicates the situation. The framework developed is used toshow that for noisy images, all edges no longer correspond to information. In this paper, varioustechniques have been presented for fusion of noisy multi-sensor images. These techniques aredeveloped for a single resolution as well as using multi-resolution decomposition. Some of thetechniques are based on modifying edge maps by filtering images, while others depend onalternate definition of information maps. Both these approaches can also be combined.Experiments show that the proposed algorithms work well for various kinds of noisy multi-sensor images
Multimodal Deep Learning for Robust RGB-D Object Recognition
Robust object recognition is a crucial ingredient of many, if not all,
real-world robotics applications. This paper leverages recent progress on
Convolutional Neural Networks (CNNs) and proposes a novel RGB-D architecture
for object recognition. Our architecture is composed of two separate CNN
processing streams - one for each modality - which are consecutively combined
with a late fusion network. We focus on learning with imperfect sensor data, a
typical problem in real-world robotics tasks. For accurate learning, we
introduce a multi-stage training methodology and two crucial ingredients for
handling depth data with CNNs. The first, an effective encoding of depth
information for CNNs that enables learning without the need for large depth
datasets. The second, a data augmentation scheme for robust learning with depth
images by corrupting them with realistic noise patterns. We present
state-of-the-art results on the RGB-D object dataset and show recognition in
challenging RGB-D real-world noisy settings.Comment: Final version submitted to IROS'2015, results unchanged,
reformulation of some text passages in abstract and introductio
- …