45,267 research outputs found
Collaborative Multi-sensor Classification via Sparsity-based Representation
In this paper, we propose a general collaborative sparse representation
framework for multi-sensor classification, which takes into account the
correlations as well as complementary information between heterogeneous sensors
simultaneously while considering joint sparsity within each sensor's
observations. We also robustify our models to deal with the presence of sparse
noise and low-rank interference signals. Specifically, we demonstrate that
incorporating the noise or interference signal as a low-rank component in our
models is essential in a multi-sensor classification problem when multiple
co-located sources/sensors simultaneously record the same physical event. We
further extend our frameworks to kernelized models which rely on sparsely
representing a test sample in terms of all the training samples in a feature
space induced by a kernel function. A fast and efficient algorithm based on
alternative direction method is proposed where its convergence to an optimal
solution is guaranteed. Extensive experiments are conducted on several real
multi-sensor data sets and results are compared with the conventional
classifiers to verify the effectiveness of the proposed methods
Generic Image Classification Approaches Excel on Face Recognition
The main finding of this work is that the standard image classification
pipeline, which consists of dictionary learning, feature encoding, spatial
pyramid pooling and linear classification, outperforms all state-of-the-art
face recognition methods on the tested benchmark datasets (we have tested on
AR, Extended Yale B, the challenging FERET, and LFW-a datasets). This
surprising and prominent result suggests that those advances in generic image
classification can be directly applied to improve face recognition systems. In
other words, face recognition may not need to be viewed as a separate object
classification problem.
While recently a large body of residual based face recognition methods focus
on developing complex dictionary learning algorithms, in this work we show that
a dictionary of randomly extracted patches (even from non-face images) can
achieve very promising results using the image classification pipeline. That
means, the choice of dictionary learning methods may not be important. Instead,
we find that learning multiple dictionaries using different low-level image
features often improve the final classification accuracy. Our proposed face
recognition approach offers the best reported results on the widely-used face
recognition benchmark datasets. In particular, on the challenging FERET and
LFW-a datasets, we improve the best reported accuracies in the literature by
about 20% and 30% respectively.Comment: 10 page
3D human action analysis and recognition through GLAC descriptor on 2D motion and static posture images
In this paper, we present an approach for identification of actions within
depth action videos. First, we process the video to get motion history images
(MHIs) and static history images (SHIs) corresponding to an action video based
on the use of 3D Motion Trail Model (3DMTM). We then characterize the action
video by extracting the Gradient Local Auto-Correlations (GLAC) features from
the SHIs and the MHIs. The two sets of features i.e., GLAC features from MHIs
and GLAC features from SHIs are concatenated to obtain a representation vector
for action. Finally, we perform the classification on all the action samples by
using the l2-regularized Collaborative Representation Classifier (l2-CRC) to
recognize different human actions in an effective way. We perform evaluation of
the proposed method on three action datasets, MSR-Action3D, DHA and UTD-MHAD.
Through experimental results, we observe that the proposed method performs
superior to other approaches.Comment: Multimed Tools Appl (2019
Collaborative Sparse Priors for Infrared Image Multi-view ATR
Feature extraction from infrared (IR) images remains a challenging task.
Learning based methods that can work on raw imagery/patches have therefore
assumed significance. We propose a novel multi-task extension of the widely
used sparse-representation-classification (SRC) method in both single and
multi-view set-ups. That is, the test sample could be a single IR image or
images from different views. When expanded in terms of a training dictionary,
the coefficient matrix in a multi-view scenario admits a sparse structure that
is not easily captured by traditional sparsity-inducing measures such as the
-row pseudo norm. To that end, we employ collaborative spike and slab
priors on the coefficient matrix, which can capture fairly general sparse
structures. Our work involves joint parameter and sparse coefficient estimation
(JPCEM) which alleviates the need to handpick prior parameters before
classification. The experimental merits of JPCEM are substantiated through
comparisons with other state-of-art methods on a challenging mid-wave IR image
(MWIR) ATR database made available by the US Army Night Vision and Electronic
Sensors Directorate.Comment: 4 pages, 3 figures, conference pape
Neither Global Nor Local: A Hierarchical Robust Subspace Clustering For Image Data
In this paper, we consider the problem of subspace clustering in presence of
contiguous noise, occlusion and disguise. We argue that self-expressive
representation of data in current state-of-the-art approaches is severely
sensitive to occlusions and complex real-world noises. To alleviate this
problem, we propose a hierarchical framework that brings robustness of local
patches-based representations and discriminant property of global
representations together. This approach consists of 1) a top-down stage, in
which the input data is subject to repeated division to smaller patches and 2)
a bottom-up stage, in which the low rank embedding of local patches in field of
view of a corresponding patch in upper level are merged on a Grassmann
manifold. This summarized information provides two key information for the
corresponding patch on the upper level: cannot-links and recommended-links.
This information is employed for computing a self-expressive representation of
each patch at upper levels using a weighted sparse group lasso optimization
problem. Numerical results on several real data sets confirm the efficiency of
our approach
Facial Expression Recognition Based on Complexity Perception Classification Algorithm
Facial expression recognition (FER) has always been a challenging issue in
computer vision. The different expressions of emotion and uncontrolled
environmental factors lead to inconsistencies in the complexity of FER and
variability of between expression categories, which is often overlooked in most
facial expression recognition systems. In order to solve this problem
effectively, we presented a simple and efficient CNN model to extract facial
features, and proposed a complexity perception classification (CPC) algorithm
for FER. The CPC algorithm divided the dataset into an easy classification
sample subspace and a complex classification sample subspace by evaluating the
complexity of facial features that are suitable for classification. The
experimental results of our proposed algorithm on Fer2013 and CK-plus datasets
demonstrated the algorithm's effectiveness and superiority over other
state-of-the-art approaches
Learning efficient sparse and low rank models
Parsimony, including sparsity and low rank, has been shown to successfully
model data in numerous machine learning and signal processing tasks.
Traditionally, such modeling approaches rely on an iterative algorithm that
minimizes an objective function with parsimony-promoting terms. The inherently
sequential structure and data-dependent complexity and latency of iterative
optimization constitute a major limitation in many applications requiring
real-time performance or involving large-scale data. Another limitation
encountered by these modeling techniques is the difficulty of their inclusion
in discriminative learning scenarios. In this work, we propose to move the
emphasis from the model to the pursuit algorithm, and develop a process-centric
view of parsimonious modeling, in which a learned deterministic
fixed-complexity pursuit process is used in lieu of iterative optimization. We
show a principled way to construct learnable pursuit process architectures for
structured sparse and robust low rank models, derived from the iteration of
proximal descent algorithms. These architectures learn to approximate the exact
parsimonious representation at a fraction of the complexity of the standard
optimization methods. We also show that appropriate training regimes allow to
naturally extend parsimonious models to discriminative settings.
State-of-the-art results are demonstrated on several challenging problems in
image and audio processing with several orders of magnitude speedup compared to
the exact optimization algorithms
A Survey on Multi-Task Learning
Multi-Task Learning (MTL) is a learning paradigm in machine learning and its
aim is to leverage useful information contained in multiple related tasks to
help improve the generalization performance of all the tasks. In this paper, we
give a survey for MTL. First, we classify different MTL algorithms into several
categories, including feature learning approach, low-rank approach, task
clustering approach, task relation learning approach, and decomposition
approach, and then discuss the characteristics of each approach. In order to
improve the performance of learning tasks further, MTL can be combined with
other learning paradigms including semi-supervised learning, active learning,
unsupervised learning, reinforcement learning, multi-view learning and
graphical models. When the number of tasks is large or the data dimensionality
is high, batch MTL models are difficult to handle this situation and online,
parallel and distributed MTL models as well as dimensionality reduction and
feature hashing are reviewed to reveal their computational and storage
advantages. Many real-world applications use MTL to boost their performance and
we review representative works. Finally, we present theoretical analyses and
discuss several future directions for MTL
Joint Maximum Purity Forest with Application to Image Super-Resolution
In this paper, we propose a novel random-forest scheme, namely Joint Maximum
Purity Forest (JMPF), for classification, clustering, and regression tasks. In
the JMPF scheme, the original feature space is transformed into a compactly
pre-clustered feature space, via a trained rotation matrix. The rotation matrix
is obtained through an iterative quantization process, where the input data
belonging to different classes are clustered to the respective vertices of the
new feature space with maximum purity. In the new feature space, orthogonal
hyperplanes, which are employed at the split-nodes of decision trees in random
forests, can tackle the clustering problems effectively. We evaluated our
proposed method on public benchmark datasets for regression and classification
tasks, and experiments showed that JMPF remarkably outperforms other
state-of-the-art random-forest-based approaches. Furthermore, we applied JMPF
to image super-resolution, because the transformed, compact features are more
discriminative to the clustering-regression scheme. Experiment results on
several public benchmark datasets also showed that the JMPF-based image
super-resolution scheme is consistently superior to recent state-of-the-art
image super-resolution algorithms.Comment: 18 pages, 7 figure
A survey of sparse representation: algorithms and applications
Sparse representation has attracted much attention from researchers in fields
of signal processing, image processing, computer vision and pattern
recognition. Sparse representation also has a good reputation in both
theoretical research and practical applications. Many different algorithms have
been proposed for sparse representation. The main purpose of this article is to
provide a comprehensive study and an updated review on sparse representation
and to supply a guidance for researchers. The taxonomy of sparse representation
methods can be studied from various viewpoints. For example, in terms of
different norm minimizations used in sparsity constraints, the methods can be
roughly categorized into five groups: sparse representation with -norm
minimization, sparse representation with -norm (0p1) minimization,
sparse representation with -norm minimization and sparse representation
with -norm minimization. In this paper, a comprehensive overview of
sparse representation is provided. The available sparse representation
algorithms can also be empirically categorized into four groups: greedy
strategy approximation, constrained optimization, proximity algorithm-based
optimization, and homotopy algorithm-based sparse representation. The
rationales of different algorithms in each category are analyzed and a wide
range of sparse representation applications are summarized, which could
sufficiently reveal the potential nature of the sparse representation theory.
Specifically, an experimentally comparative study of these sparse
representation algorithms was presented. The Matlab code used in this paper can
be available at: http://www.yongxu.org/lunwen.html.Comment: Published on IEEE Access, Vol. 3, pp. 490-530, 201
- …