1,394 research outputs found
Discriminative Scale Space Tracking
Accurate scale estimation of a target is a challenging research problem in
visual object tracking. Most state-of-the-art methods employ an exhaustive
scale search to estimate the target size. The exhaustive search strategy is
computationally expensive and struggles when encountered with large scale
variations. This paper investigates the problem of accurate and robust scale
estimation in a tracking-by-detection framework. We propose a novel scale
adaptive tracking approach by learning separate discriminative correlation
filters for translation and scale estimation. The explicit scale filter is
learned online using the target appearance sampled at a set of different
scales. Contrary to standard approaches, our method directly learns the
appearance change induced by variations in the target scale. Additionally, we
investigate strategies to reduce the computational cost of our approach.
Extensive experiments are performed on the OTB and the VOT2014 datasets.
Compared to the standard exhaustive scale search, our approach achieves a gain
of 2.5% in average overlap precision on the OTB dataset. Additionally, our
method is computationally efficient, operating at a 50% higher frame rate
compared to the exhaustive scale search. Our method obtains the top rank in
performance by outperforming 19 state-of-the-art trackers on OTB and 37
state-of-the-art trackers on VOT2014.Comment: To appear in TPAMI. This is the journal extension of the
VOT2014-winning DSST tracking metho
Robust Visual Tracking Revisited: From Correlation Filter to Template Matching
In this paper, we propose a novel matching based tracker by investigating the
relationship between template matching and the recent popular correlation
filter based trackers (CFTs). Compared to the correlation operation in CFTs, a
sophisticated similarity metric termed "mutual buddies similarity" (MBS) is
proposed to exploit the relationship of multiple reciprocal nearest neighbors
for target matching. By doing so, our tracker obtains powerful discriminative
ability on distinguishing target and background as demonstrated by both
empirical and theoretical analyses. Besides, instead of utilizing single
template with the improper updating scheme in CFTs, we design a novel online
template updating strategy named "memory filtering" (MF), which aims to select
a certain amount of representative and reliable tracking results in history to
construct the current stable and expressive template set. This scheme is
beneficial for the proposed tracker to comprehensively "understand" the target
appearance variations, "recall" some stable results. Both qualitative and
quantitative evaluations on two benchmarks suggest that the proposed tracking
method performs favorably against some recently developed CFTs and other
competitive trackers.Comment: has been published on IEEE TI
A Reverse Hierarchy Model for Predicting Eye Fixations
A number of psychological and physiological evidences suggest that early
visual attention works in a coarse-to-fine way, which lays a basis for the
reverse hierarchy theory (RHT). This theory states that attention propagates
from the top level of the visual hierarchy that processes gist and abstract
information of input, to the bottom level that processes local details.
Inspired by the theory, we develop a computational model for saliency detection
in images. First, the original image is downsampled to different scales to
constitute a pyramid. Then, saliency on each layer is obtained by image
super-resolution reconstruction from the layer above, which is defined as
unpredictability from this coarse-to-fine reconstruction. Finally, saliency on
each layer of the pyramid is fused into stochastic fixations through a
probabilistic model, where attention initiates from the top layer and
propagates downward through the pyramid. Extensive experiments on two standard
eye-tracking datasets show that the proposed method can achieve competitive
results with state-of-the-art models.Comment: CVPR 2014, 27th IEEE Conference on Computer Vision and Pattern
Recognition (CVPR). CVPR 201
- …