3,014 research outputs found
Fusion of Head and Full-Body Detectors for Multi-Object Tracking
In order to track all persons in a scene, the tracking-by-detection paradigm
has proven to be a very effective approach. Yet, relying solely on a single
detector is also a major limitation, as useful image information might be
ignored. Consequently, this work demonstrates how to fuse two detectors into a
tracking system. To obtain the trajectories, we propose to formulate tracking
as a weighted graph labeling problem, resulting in a binary quadratic program.
As such problems are NP-hard, the solution can only be approximated. Based on
the Frank-Wolfe algorithm, we present a new solver that is crucial to handle
such difficult problems. Evaluation on pedestrian tracking is provided for
multiple scenarios, showing superior results over single detector tracking and
standard QP-solvers. Finally, our tracker ranks 2nd on the MOT16 benchmark and
1st on the new MOT17 benchmark, outperforming over 90 trackers.Comment: 10 pages, 4 figures; Winner of the MOT17 challenge; CVPRW 201
Evaluating the reliability of automatically generated pedestrian and bicycle crash surrogates
Vulnerable road users (VRUs), such as pedestrians and bicyclists, are at a
higher risk of being involved in crashes with motor vehicles, and crashes
involving VRUs also are more likely to result in severe injuries or fatalities.
Signalized intersections are a major safety concern for VRUs due to their
complex and dynamic nature, highlighting the need to understand how these road
users interact with motor vehicles and deploy evidence-based countermeasures to
improve safety performance. Crashes involving VRUs are relatively infrequent,
making it difficult to understand the underlying contributing factors. An
alternative is to identify and use conflicts between VRUs and motorized
vehicles as a surrogate for safety performance. Automatically detecting these
conflicts using a video-based systems is a crucial step in developing smart
infrastructure to enhance VRU safety. The Pennsylvania Department of
Transportation conducted a study using video-based event monitoring system to
assess VRU and motor vehicle interactions at fifteen signalized intersections
across Pennsylvania to improve VRU safety performance. This research builds on
that study to assess the reliability of automatically generated surrogates in
predicting confirmed conflicts using advanced data-driven models. The surrogate
data used for analysis include automatically collectable variables such as
vehicular and VRU speeds, movements, post-encroachment time, in addition to
manually collected variables like signal states, lighting, and weather
conditions. The findings highlight the varying importance of specific
surrogates in predicting true conflicts, some being more informative than
others. The findings can assist transportation agencies to collect the right
types of data to help prioritize infrastructure investments, such as bike lanes
and crosswalks, and evaluate their effectiveness
Evidential combination of pedestrian detectors
International audienceThe importance of pedestrian detection in many applications has led to the development of many algorithms. In this paper, we address the problem of combining the outputs of several detectors. A pre-trained pedestrian detector is seen as a black box returning a set of bounding boxes with associated scores. A calibration step is first conducted to transform those scores into a probability measure. The bounding boxes are then grouped into clusters and their scores are combined. Different combination strategies using the theory of belief functions are proposed and compared to probabilistic ones. A combination rule based on triangular norms is used to deal with dependencies among detectors. More than 30 state-of-the-art detectors were combined and tested on the Caltech Pedestrian Detection Benchmark. The best combination strategy outperforms the currently best performing detector by 9% in terms of log-average miss rate
A Comprehensive Survey of Deep Learning in Remote Sensing: Theories, Tools and Challenges for the Community
In recent years, deep learning (DL), a re-branding of neural networks (NNs),
has risen to the top in numerous areas, namely computer vision (CV), speech
recognition, natural language processing, etc. Whereas remote sensing (RS)
possesses a number of unique challenges, primarily related to sensors and
applications, inevitably RS draws from many of the same theories as CV; e.g.,
statistics, fusion, and machine learning, to name a few. This means that the RS
community should be aware of, if not at the leading edge of, of advancements
like DL. Herein, we provide the most comprehensive survey of state-of-the-art
RS DL research. We also review recent new developments in the DL field that can
be used in DL for RS. Namely, we focus on theories, tools and challenges for
the RS community. Specifically, we focus on unsolved challenges and
opportunities as it relates to (i) inadequate data sets, (ii)
human-understandable solutions for modelling physical phenomena, (iii) Big
Data, (iv) non-traditional heterogeneous data sources, (v) DL architectures and
learning algorithms for spectral, spatial and temporal data, (vi) transfer
learning, (vii) an improved theoretical understanding of DL systems, (viii)
high barriers to entry, and (ix) training and optimizing the DL.Comment: 64 pages, 411 references. To appear in Journal of Applied Remote
Sensin
Asymmetric Totally-corrective Boosting for Real-time Object Detection
Real-time object detection is one of the core problems in computer vision.
The cascade boosting framework proposed by Viola and Jones has become the
standard for this problem. In this framework, the learning goal for each node
is asymmetric, which is required to achieve a high detection rate and a
moderate false positive rate. We develop new boosting algorithms to address
this asymmetric learning problem. We show that our methods explicitly optimize
asymmetric loss objectives in a totally corrective fashion. The methods are
totally corrective in the sense that the coefficients of all selected weak
classifiers are updated at each iteration. In contract, conventional boosting
like AdaBoost is stage-wise in that only the current weak classifier's
coefficient is updated. At the heart of the totally corrective boosting is the
column generation technique. Experiments on face detection show that our
methods outperform the state-of-the-art asymmetric boosting methods.Comment: 14 pages, published in Asian Conf. Computer Vision 201
Multi-sensor fusion based on multiple classifier systems for human activity identification
Multimodal sensors in healthcare applications have been increasingly researched because it facilitates automatic and comprehensive monitoring of human behaviors, high-intensity sports management, energy expenditure estimation, and postural detection. Recent studies have shown the importance of multi-sensor fusion to achieve robustness, high-performance generalization, provide diversity and tackle challenging issue that maybe difficult with single sensor values. The aim of this study is to propose an innovative multi-sensor fusion framework to improve human activity detection performances and reduce misrecognition rate. The study proposes a multi-view ensemble algorithm to integrate predicted values of different motion sensors. To this end, computationally efficient classification algorithms such as decision tree, logistic regression and k-Nearest Neighbors were used to implement diverse, flexible and dynamic human activity detection systems. To provide compact feature vector representation, we studied hybrid bio-inspired evolutionary search algorithm and correlation-based feature selection method and evaluate their impact on extracted feature vectors from individual sensor modality. Furthermore, we utilized Synthetic Over-sampling minority Techniques (SMOTE) algorithm to reduce the impact of class imbalance and improve performance results. With the above methods, this paper provides unified framework to resolve major challenges in human activity identification. The performance results obtained using two publicly available datasets showed significant improvement over baseline methods in the detection of specific activity details and reduced error rate. The performance results of our evaluation showed 3% to 24% improvement in accuracy, recall, precision, F-measure and detection ability (AUC) compared to single sensors and feature-level fusion. The benefit of the proposed multi-sensor fusion is the ability to utilize distinct feature characteristics of individual sensor and multiple classifier systems to improve recognition accuracy. In addition, the study suggests a promising potential of hybrid feature selection approach, diversity-based multiple classifier systems to improve mobile and wearable sensor-based human activity detection and health monitoring system. - 2019, The Author(s).This research is supported by University of Malaya BKP Special Grant no vote BKS006-2018.Scopu
- …