397,959 research outputs found
Recommended from our members
Recognition of human interactions with vehicles using 3-D models and dynamic context
textThis dissertation describes two distinctive methods for human-vehicle interaction recognition: one for ground level videos and the other for aerial videos. For ground level videos, this dissertation presents a novel methodology which is able to estimate a detailed status of a scene involving multiple humans and vehicles. The system tracks their configuration even when they are performing complex interactions with severe occlusion such as when four persons are exiting a car together. The motivation is to identify the 3-D states of vehicles (e.g. status of doors), their relations with persons, which is necessary to analyze complex human-vehicle interactions (e.g. breaking into or stealing a vehicle), and the motion of humans and car doors to detect atomic human-vehicle interactions. A probabilistic algorithm has been designed to track humans and analyze their dynamic relationships with vehicles using a dynamic context. We have focused on two ideas. One is that many simple events can be detected based on a low-level analysis, and these detected events must contextually meet with human/vehicle status tracking results. The other is that the motion clue interferes with states in the current and future frames, and analyzing the motion is critical to detect such simple events. Our approach updates the probability of a person (or a vehicle) having a particular state based on these basic observed events. The probabilistic inference is made for the tracking process to match event-based evidence and motion-based evidence. For aerial videos, the object resolution is low, the visual cues are vague, and the detection and tracking of objects is less reliable as a consequence. Any method that requires accurate tracking of objects or the exact matching of event definition are better avoided. To address these issues, we present a temporal logic based approach which does not require training from event examples. At the low-level, we employ dynamic programming to perform fast model fitting between the tracked vehicle and the rendered 3-D vehicle models. At the semantic-level, given the localized event region of interest (ROI), we verify the time series of human-vehicle relationships with the pre-specified event definitions in a piecewise fashion. With special interest in recognizing a person getting into and out of a vehicle, we have tested our method on a subset of the VIRAT Aerial Video dataset and achieved superior results.Electrical and Computer Engineerin
A Model-based Approach for Fast Vehicle Detection in Continuously Streamed Urban LIDAR Point Clouds
Detection of vehicles in crowded 3-D urban scenes is a challenging problem in many computer vision related research fields, such as robot perception, autonomous driving, self-localization, and mapping. In this paper we present a model-based approach to solve the recognition problem from 3-D range data. In particular, we aim to detect and recognize vehicles from continuously streamed LIDAR point cloud sequences of a rotating multi-beam laser scanner. The end-to-end pipeline of our framework working on the raw streams of 3-D urban laser data consists of three steps 1) producing distinct groups of points which represent different urban objects 2) extracting reliable 3-D shape descriptors specifically designed for vehicles, considering the need for fast processing speed 3) executing binary classification on the extracted descriptors in order to perform vehicle detection. The extraction of our efficient shape
descriptors provides a significant speedup with and increased detection
accuracy compared to a PCA based 3-D bounding box fitting method
used as baseline
Vehicles Recognition Using Fuzzy Descriptors of Image Segments
In this paper a vision-based vehicles recognition method is presented.
Proposed method uses fuzzy description of image segments for automatic
recognition of vehicles recorded in image data. The description takes into
account selected geometrical properties and shape coefficients determined for
segments of reference image (vehicle model). The proposed method was
implemented using reasoning system with fuzzy rules. A vehicles recognition
algorithm was developed based on the fuzzy rules describing shape and
arrangement of the image segments that correspond to visible parts of a
vehicle. An extension of the algorithm with set of fuzzy rules defined for
different reference images (and various vehicle shapes) enables vehicles
classification in traffic scenes. The devised method is suitable for
application in video sensors for road traffic control and surveillance systems.Comment: The final publication is available at http://www.springerlink.co
A Robust Real-Time Automatic License Plate Recognition Based on the YOLO Detector
Automatic License Plate Recognition (ALPR) has been a frequent topic of
research due to many practical applications. However, many of the current
solutions are still not robust in real-world situations, commonly depending on
many constraints. This paper presents a robust and efficient ALPR system based
on the state-of-the-art YOLO object detector. The Convolutional Neural Networks
(CNNs) are trained and fine-tuned for each ALPR stage so that they are robust
under different conditions (e.g., variations in camera, lighting, and
background). Specially for character segmentation and recognition, we design a
two-stage approach employing simple data augmentation tricks such as inverted
License Plates (LPs) and flipped characters. The resulting ALPR approach
achieved impressive results in two datasets. First, in the SSIG dataset,
composed of 2,000 frames from 101 vehicle videos, our system achieved a
recognition rate of 93.53% and 47 Frames Per Second (FPS), performing better
than both Sighthound and OpenALPR commercial systems (89.80% and 93.03%,
respectively) and considerably outperforming previous results (81.80%). Second,
targeting a more realistic scenario, we introduce a larger public dataset,
called UFPR-ALPR dataset, designed to ALPR. This dataset contains 150 videos
and 4,500 frames captured when both camera and vehicles are moving and also
contains different types of vehicles (cars, motorcycles, buses and trucks). In
our proposed dataset, the trial versions of commercial systems achieved
recognition rates below 70%. On the other hand, our system performed better,
with recognition rate of 78.33% and 35 FPS.Comment: Accepted for presentation at the International Joint Conference on
Neural Networks (IJCNN) 201
Ego-motion and Surrounding Vehicle State Estimation Using a Monocular Camera
Understanding ego-motion and surrounding vehicle state is essential to enable
automated driving and advanced driving assistance technologies. Typical
approaches to solve this problem use fusion of multiple sensors such as LiDAR,
camera, and radar to recognize surrounding vehicle state, including position,
velocity, and orientation. Such sensing modalities are overly complex and
costly for production of personal use vehicles. In this paper, we propose a
novel machine learning method to estimate ego-motion and surrounding vehicle
state using a single monocular camera. Our approach is based on a combination
of three deep neural networks to estimate the 3D vehicle bounding box, depth,
and optical flow from a sequence of images. The main contribution of this paper
is a new framework and algorithm that integrates these three networks in order
to estimate the ego-motion and surrounding vehicle state. To realize more
accurate 3D position estimation, we address ground plane correction in
real-time. The efficacy of the proposed method is demonstrated through
experimental evaluations that compare our results to ground truth data
available from other sensors including Can-Bus and LiDAR
- …