190 research outputs found

    Forward Vehicle Collision Warning Based on Quick Camera Calibration

    Full text link
    Forward Vehicle Collision Warning (FCW) is one of the most important functions for autonomous vehicles. In this procedure, vehicle detection and distance measurement are core components, requiring accurate localization and estimation. In this paper, we propose a simple but efficient forward vehicle collision warning framework by aggregating monocular distance measurement and precise vehicle detection. In order to obtain forward vehicle distance, a quick camera calibration method which only needs three physical points to calibrate related camera parameters is utilized. As for the forward vehicle detection, a multi-scale detection algorithm that regards the result of calibration as distance priori is proposed to improve the precision. Intensive experiments are conducted in our established real scene dataset and the results have demonstrated the effectiveness of the proposed framework

    Multi-view Vehicle Detection based on Part Model with Active Learning

    Full text link
    © 2018 IEEE. Nowadays, most ofthe vehicle detection methods aim to detect only single-view vehicles, and the performance is easily affected by partial occlusion. Therefore, a novel multi-view vehicle detection system is proposed to solve the problem of partial occlusion. The proposed system is divided into two steps: Background filtering and part model. Background filtering step is used to filter out trees, sky and other road background objects. In the part model step, each of the part models is trained by samples collected by using the proposed active learning algorithm. This paper validates the performance of the background filtering method and the part model algorithm in multi-view car detection. The performance of the proposed method outperforms previously proposed methods

    3D Object Representations for Recognition.

    Full text link
    Object recognition from images is a longstanding and challenging problem in computer vision. The main challenge is that the appearance of objects in images is affected by a number of factors, such as illumination, scale, camera viewpoint, intra-class variability, occlusion, truncation, and so on. How to handle all these factors in object recognition is still an open problem. In this dissertation, I present my efforts in building 3D object representations for object recognition. Compared to 2D appearance based object representations, 3D object representations can capture the 3D nature of objects and better handle viewpoint variation, occlusion and truncation in object recognition. I introduce three new 3D object representations: the 3D aspect part representation, the 3D aspectlet representation and the 3D voxel pattern representation. These representations are built to handle different challenging factors in object recognition. The 3D aspect part representation is able to capture the appearance change of object categories due to viewpoint transformation. The 3D aspectlet representation and the 3D voxel pattern representation are designed to handle occlusions between objects in addition to viewpoint change. Based on these representations, we propose new object recognition methods and conduct experiments on benchmark datasets to verify the advantages of our methods. Furthermore, we introduce, PASCAL3D+, a new large scale dataset for 3D object recognition by aligning objects in images with 3D CAD models. We also propose two novel methods to tackle object co-detection and multiview object tracking using our 3D aspect part representation, and a novel Convolutional Neural Network-based approach for object detection using our 3D voxel pattern representation. In order to track multiple objects in videos, we introduce a new online multi-object tracking framework based on Markov Decision Processes. Lastly, I conclude the dissertation and discuss future steps for 3D object recognition.PhDElectrical Engineering: SystemsUniversity of Michigan, Horace H. Rackham School of Graduate Studieshttp://deepblue.lib.umich.edu/bitstream/2027.42/120836/1/yuxiang_1.pd

    Vehicle pose estimation for vehicle detection and tracking based on road direction

    Get PDF
    Vehicle has several types and each of them has different color, size, and shape. The appearance of vehicle also changes if viewed from different viewpoint of traffic surveillance camera. This situation can create many possibilities of vehicle poses. However, the one in common, vehicle pose usually follows road direction. Therefore, this research proposes a method to estimate the pose of vehicle for vehicle detection and tracking based on road direction. Vehicle training data are generated from 3D vehicle models in four-pair orientation categories. Histogram of Oriented Gradients (HOG) and Linear-Support Vector Machine (Linear-SVM) are used to build vehicle detectors from the data. Road area is extracted from traffic surveillance image to localize the detection area. The pose of vehicle which estimated based on road direction will be used to select a suitable vehicle detector for vehicle detection process. To obtain the final vehicle object, vehicle line checking method is applied to the vehicle detection result. Finally, vehicle tracking is performed to give label on each vehicle. The test conducted on various viewpoints of traffic surveillance camera shows that the method effectively detects and tracks vehicle by estimating the pose of vehicle. Performance evaluation of the proposed method shows 0.9170 of accuracy and 0.9161 of balance accuracy (BAC)

    Vehicle pose estimation for vehicle detection and tracking based on road direction

    Get PDF

    Conditional Random Fields for Multi-Camera Object Detection

    Get PDF
    We formulate a model for multi-class object detection in a multi-camera environment. From our knowledge, this is the first time that this problem is addressed taken into account different object classes simultaneously. Given several images of the scene taken from different angles, our system estimates the ground plane location of the objects from the output of several object detectors applied at each viewpoint. We cast the problem as an energy minimization modeled with a Conditional Random Field (CRF). Instead of predicting the presence of an object at each image location independently, we simultaneously predict the labeling of the entire scene. Our CRF is able to take into account occlusions between objects and contextual constraints among them. We propose an effective iterative strategy that renders tractable the underlying optimization problem, and learn the parameters of the model with the max-margin paradigm. We evaluate the performance of our model on several challenging multi-camera pedestrian detection datasets namely PETS 2009 and EPFL terrace sequence. We also introduce a new dataset in which multiple classes of objects appear simultaneously in the scene. It is here where we show that our method effectively handles occlusions in the multi-class case
    • …
    corecore