4 research outputs found
Image segmentation with superpixel-based covariance descriptors in low-rank representation
This paper investigates the problem of image segmentation using superpixels.
We propose two approaches to enhance the discriminative ability of the
superpixel's covariance descriptors. In the first one, we employ the
Log-Euclidean distance as the metric on the covariance manifolds, and then use
the RBF kernel to measure the similarities between covariance descriptors. The
second method is focused on extracting the subspace structure of the set of
covariance descriptors by extending a low rank representation algorithm on to
the covariance manifolds. Experiments are carried out with the Berkly
Segmentation Dataset, and compared with the state-of-the-art segmentation
algorithms, both methods are competitive.Comment: 7 pages, 2 figures, 1 tabl
Localized LRR on Grassmann Manifolds: An Extrinsic View
Subspace data representation has recently become a common practice in many
computer vision tasks. It demands generalizing classical machine learning
algorithms for subspace data. Low-Rank Representation (LRR) is one of the most
successful models for clustering vectorial data according to their subspace
structures. This paper explores the possibility of extending LRR for subspace
data on Grassmann manifolds. Rather than directly embedding the Grassmann
manifolds into the symmetric matrix space, an extrinsic view is taken to build
the LRR self-representation in the local area of the tangent space at each
Grassmannian point, resulting in a localized LRR method on Grassmann manifolds.
A novel algorithm for solving the proposed model is investigated and
implemented. The performance of the new clustering algorithm is assessed
through experiments on several real-world datasets including MNIST handwritten
digits, ballet video clips, SKIG action clips, DynTex++ dataset and highway
traffic video clips. The experimental results show the new method outperforms a
number of state-of-the-art clustering methodsComment: IEEE Transactions on Circuits and Systems for Video Technology with
Minor Revisions. arXiv admin note: text overlap with arXiv:1504.0180
Constructing the L2-Graph for Robust Subspace Learning and Subspace Clustering
Under the framework of graph-based learning, the key to robust subspace
clustering and subspace learning is to obtain a good similarity graph that
eliminates the effects of errors and retains only connections between the data
points from the same subspace (i.e., intra-subspace data points). Recent works
achieve good performance by modeling errors into their objective functions to
remove the errors from the inputs. However, these approaches face the
limitations that the structure of errors should be known prior and a complex
convex problem must be solved. In this paper, we present a novel method to
eliminate the effects of the errors from the projection space (representation)
rather than from the input space. We first prove that -, -,
-, and nuclear-norm based linear projection spaces share the
property of Intra-subspace Projection Dominance (IPD), i.e., the coefficients
over intra-subspace data points are larger than those over inter-subspace data
points. Based on this property, we introduce a method to construct a sparse
similarity graph, called L2-Graph. The subspace clustering and subspace
learning algorithms are developed upon L2-Graph. Experiments show that L2-Graph
algorithms outperform the state-of-the-art methods for feature extraction,
image clustering, and motion segmentation in terms of accuracy, robustness, and
time efficiency
A Forward Backward Greedy approach for Sparse Multiscale Learning
Multiscale Models are known to be successful in uncovering and analyzing the
structures in data at different resolutions. In the current work we propose a
feature driven Reproducing Kernel Hilbert space (RKHS), for which the
associated kernel has a weighted multiscale structure. For generating
approximations in this space, we provide a practical forward-backward algorithm
that is shown to greedily construct a set of basis functions having a
multiscale structure, while also creating sparse representations from the given
data set, making representations and predictions very efficient. We provide a
detailed analysis of the algorithm including recommendations for selecting
algorithmic hyper-parameters and estimating probabilistic rates of convergence
at individual scales. Then we extend this analysis to multiscale setting,
studying the effects of finite scale truncation and quality of solution in the
inherent RKHS. In the last section, we analyze the performance of the approach
on a variety of simulation and real data sets, thereby justifying the
efficiency claims in terms of model quality and data reduction