1,987 research outputs found
A Study of Exploiting Objectness for Robust Online Object Tracking
Tracking is a fundamental problem in many computer vision applications. Despite the progress over the last decade, there still exist many challenges especially when the problem is posed in real world scenarios (e.g., cluttered background, occluded objects). Among them drifting has been widely observed to be a problem common to the class of online tracking algorithms - i.e., when challenges such as occlusion or nonlinear deformation of the object occurs, the tracker might lose the target completely in subsequent frames in an image sequence. In this work, we propose to exploit the objectness to partially alleviate the drifting problem with the class of online object tracking and verify the effectiveness of this idea by extensive experimental results. More specifically, a recently developed objectness measure was incorporated into Incremental Learning for Visual Tracking (IVT) algorithm in a principled way. We have come up with a strategy of reinitializing the training samples in the proposed approach to improve the robustness of online tracking. Experimental results show that using objectness measure does help to alleviate its drift to background for certain challenging sequences
Single and multiple object tracking using a multi-feature joint sparse representation
In this paper, we propose a tracking algorithm based on a multi-feature joint sparse representation. The templates for the sparse representation can include pixel values, textures, and edges. In the multi-feature joint optimization, noise or occlusion is dealt with using a set of trivial templates. A sparse weight constraint is introduced to dynamically select the relevant templates from the full set of templates. A variance ratio measure is adopted to adaptively adjust the weights of different features. The multi-feature template set is updated adaptively. We further propose an algorithm for tracking multi-objects with occlusion handling based on the multi-feature joint sparse reconstruction. The observation model based on sparse reconstruction automatically focuses on the visible parts of an occluded object by using the information in the trivial templates. The multi-object tracking is simplified into a joint Bayesian inference. The experimental results show the superiority of our algorithm over several state-of-the-art tracking algorithms
A Unified Framework for Compositional Fitting of Active Appearance Models
Active Appearance Models (AAMs) are one of the most popular and
well-established techniques for modeling deformable objects in computer vision.
In this paper, we study the problem of fitting AAMs using Compositional
Gradient Descent (CGD) algorithms. We present a unified and complete view of
these algorithms and classify them with respect to three main characteristics:
i) cost function; ii) type of composition; and iii) optimization method.
Furthermore, we extend the previous view by: a) proposing a novel Bayesian cost
function that can be interpreted as a general probabilistic formulation of the
well-known project-out loss; b) introducing two new types of composition,
asymmetric and bidirectional, that combine the gradients of both image and
appearance model to derive better conver- gent and more robust CGD algorithms;
and c) providing new valuable insights into existent CGD algorithms by
reinterpreting them as direct applications of the Schur complement and the
Wiberg method. Finally, in order to encourage open research and facilitate
future comparisons with our work, we make the implementa- tion of the
algorithms studied in this paper publicly available as part of the Menpo
Project.Comment: 39 page
Efficient illumination independent appearance-based face tracking
One of the major challenges that visual tracking algorithms face nowadays is being
able to cope with changes in the appearance of the target during tracking. Linear
subspace models have been extensively studied and are possibly the most popular
way of modelling target appearance. We introduce a linear subspace representation
in which the appearance of a face is represented by the addition of two approxi-
mately independent linear subspaces modelling facial expressions and illumination
respectively. This model is more compact than previous bilinear or multilinear ap-
proaches. The independence assumption notably simplifies system training. We only
require two image sequences. One facial expression is subject to all possible illumina-
tions in one sequence and the face adopts all facial expressions under one particular
illumination in the other. This simple model enables us to train the system with
no manual intervention. We also revisit the problem of efficiently fitting a linear
subspace-based model to a target image and introduce an additive procedure for
solving this problem. We prove that Matthews and Baker’s Inverse Compositional
Approach makes a smoothness assumption on the subspace basis that is equiva-
lent to Hager and Belhumeur’s, which worsens convergence. Our approach differs
from Hager and Belhumeur’s additive and Matthews and Baker’s compositional ap-
proaches in that we make no smoothness assumptions on the subspace basis. In the
experiments conducted we show that the model introduced accurately represents
the appearance variations caused by illumination changes and facial expressions.
We also verify experimentally that our fitting procedure is more accurate and has
better convergence rate than the other related approaches, albeit at the expense of
a slight increase in computational cost. Our approach can be used for tracking a
human face at standard video frame rates on an average personal computer
- …