874 research outputs found
KCRC-LCD: Discriminative Kernel Collaborative Representation with Locality Constrained Dictionary for Visual Categorization
We consider the image classification problem via kernel collaborative
representation classification with locality constrained dictionary (KCRC-LCD).
Specifically, we propose a kernel collaborative representation classification
(KCRC) approach in which kernel method is used to improve the discrimination
ability of collaborative representation classification (CRC). We then measure
the similarities between the query and atoms in the global dictionary in order
to construct a locality constrained dictionary (LCD) for KCRC. In addition, we
discuss several similarity measure approaches in LCD and further present a
simple yet effective unified similarity measure whose superiority is validated
in experiments. There are several appealing aspects associated with LCD. First,
LCD can be nicely incorporated under the framework of KCRC. The LCD similarity
measure can be kernelized under KCRC, which theoretically links CRC and LCD
under the kernel method. Second, KCRC-LCD becomes more scalable to both the
training set size and the feature dimension. Example shows that KCRC is able to
perfectly classify data with certain distribution, while conventional CRC fails
completely. Comprehensive experiments on many public datasets also show that
KCRC-LCD is a robust discriminative classifier with both excellent performance
and good scalability, being comparable or outperforming many other
state-of-the-art approaches
Robust Principal Component Analysis?
This paper is about a curious phenomenon. Suppose we have a data matrix,
which is the superposition of a low-rank component and a sparse component. Can
we recover each component individually? We prove that under some suitable
assumptions, it is possible to recover both the low-rank and the sparse
components exactly by solving a very convenient convex program called Principal
Component Pursuit; among all feasible decompositions, simply minimize a
weighted combination of the nuclear norm and of the L1 norm. This suggests the
possibility of a principled approach to robust principal component analysis
since our methodology and results assert that one can recover the principal
components of a data matrix even though a positive fraction of its entries are
arbitrarily corrupted. This extends to the situation where a fraction of the
entries are missing as well. We discuss an algorithm for solving this
optimization problem, and present applications in the area of video
surveillance, where our methodology allows for the detection of objects in a
cluttered background, and in the area of face recognition, where it offers a
principled way of removing shadows and specularities in images of faces
FACE RECOGNITION AND VERIFICATION IN UNCONSTRAINED ENVIRIONMENTS
Face recognition has been a long standing problem in computer vision. General
face recognition is challenging because of large appearance variability due to
factors including pose, ambient lighting, expression, size of the face, age, and distance
from the camera, etc. There are very accurate techniques to perform face
recognition in controlled environments, especially when large numbers of samples
are available for each face (individual). However, face identification under uncontrolled(
unconstrained) environments or with limited training data is still an unsolved
problem. There are two face recognition tasks: face identification (who is who in
a probe face set, given a gallery face set) and face verification (same or not, given
two faces). In this work, we study both face identification and verification in unconstrained
environments.
Firstly, we propose a face verification framework that combines Partial Least
Squares (PLS) and the One-Shot similarity model[1]. The idea is to describe a
face with a large feature set combining shape, texture and color information. PLS
regression is applied to perform multi-channel feature weighting on this large feature
set. Finally the PLS regression is used to compute the similarity score of an image
pair by One-Shot learning (using a fixed negative set).
Secondly, we study face identification with image sets, where the gallery and
probe are sets of face images of an individual. We model a face set by its covariance
matrix (COV) which is a natural 2nd-order statistic of a sample set.By exploring an
efficient metric for the SPD matrices, i.e., Log-Euclidean Distance (LED), we derive
a kernel function that explicitly maps the covariance matrix from the Riemannian
manifold to Euclidean space. Then, discriminative learning is performed on the
COV manifold: the learning aims to maximize the between-class COV distance and
minimize the within-class COV distance.
Sparse representation and dictionary learning have been widely used in face
recognition, especially when large numbers of samples are available for each face
(individual). Sparse coding is promising since it provides a more stable and discriminative
face representation. In the last part of our work, we explore sparse
coding and dictionary learning for face verification application. More specifically,
in one approach, we apply sparse representations to face verification in two ways
via a fix reference set as dictionary. In the other approach, we propose a dictionary
learning framework with explicit pairwise constraints, which unifies the discriminative
dictionary learning for pair matching (face verification) and classification (face
recognition) problems
Sparse Variation Dictionary Learning for Face Recognition with a Single Training Sample per Person
Face recognition (FR) with a single training sample per person (STSPP) is a very challenging problem due to the lack of information to predict the variations in the query sample. Sparse representation based classification has shown interesting results in robust FR, however, its performance will deteriorate much for FR with STSPP. To address this issue, in this paper we learn a sparse variation dictionary from a generic training set to improve the query sample representation by STSPP. Instead of learning from the generic training set independently w.r.t. the gallery set, the proposed sparse variation dictionary learning (SVDL) method is adaptive to the gallery set by jointly learning a projection to connect the generic training set with the gallery set. The learnt sparse variation dictionary can be easily integrated into the framework of sparse representation based classification so that various variations in face images, including illumination, expression, occlusion, pose, etc., can be better handled. Experiments on the large-scale CMU Multi-PIE, FRGC and LFW databases demonstrate the promising performance of SVDL on FR with STSPP.Department of ComputingRefereed conference pape
- …