8,921 research outputs found
Graph Connectivity in Noisy Sparse Subspace Clustering
Subspace clustering is the problem of clustering data points into a union of
low-dimensional linear/affine subspaces. It is the mathematical abstraction of
many important problems in computer vision, image processing and machine
learning. A line of recent work (4, 19, 24, 20) provided strong theoretical
guarantee for sparse subspace clustering (4), the state-of-the-art algorithm
for subspace clustering, on both noiseless and noisy data sets. It was shown
that under mild conditions, with high probability no two points from different
subspaces are clustered together. Such guarantee, however, is not sufficient
for the clustering to be correct, due to the notorious "graph connectivity
problem" (15). In this paper, we investigate the graph connectivity problem for
noisy sparse subspace clustering and show that a simple post-processing
procedure is capable of delivering consistent clustering under certain "general
position" or "restricted eigenvalue" assumptions. We also show that our
condition is almost tight with adversarial noise perturbation by constructing a
counter-example. These results provide the first exact clustering guarantee of
noisy SSC for subspaces of dimension greater then 3.Comment: 14 pages. To appear in The 19th International Conference on
Artificial Intelligence and Statistics, held at Cadiz, Spain in 201
Recommended from our members
Tensor LRR and sparse coding-based subspace clustering
Subspace clustering groups a set of samples from a union of several linear subspaces into clusters, so that the samples in the same cluster are drawn from the same linear subspace. In the majority of the existing work on subspace clustering, clusters are built based on feature information, while sample correlations in their original spatial structure are simply ignored. Besides, original high-dimensional feature vector contains noisy/redundant information, and the time complexity grows exponentially with the number of dimensions. To address these issues, we propose a tensor low-rank representation (TLRR) and sparse coding-based (TLRRSC) subspace clustering method by simultaneously considering feature information and spatial structures. TLRR seeks the lowest rank representation over original spatial structures along all spatial directions. Sparse coding learns a dictionary along feature spaces, so that each sample can be represented by a few atoms of the learned dictionary. The affinity matrix used for spectral clustering is built from the joint similarities in both spatial and feature spaces. TLRRSC can well capture the global structure and inherent feature information of data, and provide a robust subspace segmentation from corrupted data. Experimental results on both synthetic and real-world data sets show that TLRRSC outperforms several established state-of-the-art methods
A Study on Clustering for Clustering Based Image De-Noising
In this paper, the problem of de-noising of an image contaminated with
Additive White Gaussian Noise (AWGN) is studied. This subject is an open
problem in signal processing for more than 50 years. Local methods suggested in
recent years, have obtained better results than global methods. However by more
intelligent training in such a way that first, important data is more effective
for training, second, clustering in such way that training blocks lie in
low-rank subspaces, we can design a dictionary applicable for image de-noising
and obtain results near the state of the art local methods. In the present
paper, we suggest a method based on global clustering of image constructing
blocks. As the type of clustering plays an important role in clustering-based
de-noising methods, we address two questions about the clustering. The first,
which parts of the data should be considered for clustering? and the second,
what data clustering method is suitable for de-noising.? Then clustering is
exploited to learn an over complete dictionary. By obtaining sparse
decomposition of the noisy image blocks in terms of the dictionary atoms, the
de-noised version is achieved. In addition to our framework, 7 popular
dictionary learning methods are simulated and compared. The results are
compared based on two major factors: (1) de-noising performance and (2)
execution time. Experimental results show that our dictionary learning
framework outperforms its competitors in terms of both factors.Comment: 9 pages, 8 figures, Journal of Information Systems and
Telecommunications (JIST
- …