10,203 research outputs found
Multiple Object Tracking in Urban Traffic Scenes with a Multiclass Object Detector
Multiple object tracking (MOT) in urban traffic aims to produce the
trajectories of the different road users that move across the field of view
with different directions and speeds and that can have varying appearances and
sizes. Occlusions and interactions among the different objects are expected and
common due to the nature of urban road traffic. In this work, a tracking
framework employing classification label information from a deep learning
detection approach is used for associating the different objects, in addition
to object position and appearances. We want to investigate the performance of a
modern multiclass object detector for the MOT task in traffic scenes. Results
show that the object labels improve tracking performance, but that the output
of object detectors are not always reliable.Comment: 13th International Symposium on Visual Computing (ISVC
Fast traffic sign recognition using color segmentation and deep convolutional networks
The use of Computer Vision techniques for the automatic
recognition of road signs is fundamental for the development of intelli-
gent vehicles and advanced driver assistance systems. In this paper, we
describe a procedure based on color segmentation, Histogram of Ori-
ented Gradients (HOG), and Convolutional Neural Networks (CNN) for
detecting and classifying road signs. Detection is speeded up by a pre-
processing step to reduce the search space, while classication is carried
out by using a Deep Learning technique. A quantitative evaluation of the
proposed approach has been conducted on the well-known German Traf-
c Sign data set and on the novel Data set of Italian Trac Signs (DITS),
which is publicly available and contains challenging sequences captured
in adverse weather conditions and in an urban scenario at night-time.
Experimental results demonstrate the eectiveness of the proposed ap-
proach in terms of both classication accuracy and computational speed
The Cityscapes Dataset for Semantic Urban Scene Understanding
Visual understanding of complex urban street scenes is an enabling factor for
a wide range of applications. Object detection has benefited enormously from
large-scale datasets, especially in the context of deep learning. For semantic
urban scene understanding, however, no current dataset adequately captures the
complexity of real-world urban scenes.
To address this, we introduce Cityscapes, a benchmark suite and large-scale
dataset to train and test approaches for pixel-level and instance-level
semantic labeling. Cityscapes is comprised of a large, diverse set of stereo
video sequences recorded in streets from 50 different cities. 5000 of these
images have high quality pixel-level annotations; 20000 additional images have
coarse annotations to enable methods that leverage large volumes of
weakly-labeled data. Crucially, our effort exceeds previous attempts in terms
of dataset size, annotation richness, scene variability, and complexity. Our
accompanying empirical study provides an in-depth analysis of the dataset
characteristics, as well as a performance evaluation of several
state-of-the-art approaches based on our benchmark.Comment: Includes supplemental materia
Model Adaptation with Synthetic and Real Data for Semantic Dense Foggy Scene Understanding
This work addresses the problem of semantic scene understanding under dense
fog. Although considerable progress has been made in semantic scene
understanding, it is mainly related to clear-weather scenes. Extending
recognition methods to adverse weather conditions such as fog is crucial for
outdoor applications. In this paper, we propose a novel method, named
Curriculum Model Adaptation (CMAda), which gradually adapts a semantic
segmentation model from light synthetic fog to dense real fog in multiple
steps, using both synthetic and real foggy data. In addition, we present three
other main stand-alone contributions: 1) a novel method to add synthetic fog to
real, clear-weather scenes using semantic input; 2) a new fog density
estimator; 3) the Foggy Zurich dataset comprising real foggy images,
with pixel-level semantic annotations for images with dense fog. Our
experiments show that 1) our fog simulation slightly outperforms a
state-of-the-art competing simulation with respect to the task of semantic
foggy scene understanding (SFSU); 2) CMAda improves the performance of
state-of-the-art models for SFSU significantly by leveraging unlabeled real
foggy data. The datasets and code are publicly available.Comment: final version, ECCV 201
- …