1,046 research outputs found
Hyperspectral Unmixing Overview: Geometrical, Statistical, and Sparse Regression-Based Approaches
Imaging spectrometers measure electromagnetic energy scattered in their
instantaneous field view in hundreds or thousands of spectral channels with
higher spectral resolution than multispectral cameras. Imaging spectrometers
are therefore often referred to as hyperspectral cameras (HSCs). Higher
spectral resolution enables material identification via spectroscopic analysis,
which facilitates countless applications that require identifying materials in
scenarios unsuitable for classical spectroscopic analysis. Due to low spatial
resolution of HSCs, microscopic material mixing, and multiple scattering,
spectra measured by HSCs are mixtures of spectra of materials in a scene. Thus,
accurate estimation requires unmixing. Pixels are assumed to be mixtures of a
few materials, called endmembers. Unmixing involves estimating all or some of:
the number of endmembers, their spectral signatures, and their abundances at
each pixel. Unmixing is a challenging, ill-posed inverse problem because of
model inaccuracies, observation noise, environmental conditions, endmember
variability, and data set size. Researchers have devised and investigated many
models searching for robust, stable, tractable, and accurate unmixing
algorithms. This paper presents an overview of unmixing methods from the time
of Keshava and Mustard's unmixing tutorial [1] to the present. Mixing models
are first discussed. Signal-subspace, geometrical, statistical, sparsity-based,
and spatial-contextual unmixing algorithms are described. Mathematical problems
and potential solutions are described. Algorithm characteristics are
illustrated experimentally.Comment: This work has been accepted for publication in IEEE Journal of
Selected Topics in Applied Earth Observations and Remote Sensin
Proceedings of the second "international Traveling Workshop on Interactions between Sparse models and Technology" (iTWIST'14)
The implicit objective of the biennial "international - Traveling Workshop on
Interactions between Sparse models and Technology" (iTWIST) is to foster
collaboration between international scientific teams by disseminating ideas
through both specific oral/poster presentations and free discussions. For its
second edition, the iTWIST workshop took place in the medieval and picturesque
town of Namur in Belgium, from Wednesday August 27th till Friday August 29th,
2014. The workshop was conveniently located in "The Arsenal" building within
walking distance of both hotels and town center. iTWIST'14 has gathered about
70 international participants and has featured 9 invited talks, 10 oral
presentations, and 14 posters on the following themes, all related to the
theory, application and generalization of the "sparsity paradigm":
Sparsity-driven data sensing and processing; Union of low dimensional
subspaces; Beyond linear and convex inverse problem; Matrix/manifold/graph
sensing/processing; Blind inverse problems and dictionary learning; Sparsity
and computational neuroscience; Information theory, geometry and randomness;
Complexity/accuracy tradeoffs in numerical methods; Sparsity? What's next?;
Sparse machine learning and inference.Comment: 69 pages, 24 extended abstracts, iTWIST'14 website:
http://sites.google.com/site/itwist1
Robust Principal Component Analysis on Graphs
Principal Component Analysis (PCA) is the most widely used tool for linear
dimensionality reduction and clustering. Still it is highly sensitive to
outliers and does not scale well with respect to the number of data samples.
Robust PCA solves the first issue with a sparse penalty term. The second issue
can be handled with the matrix factorization model, which is however
non-convex. Besides, PCA based clustering can also be enhanced by using a graph
of data similarity. In this article, we introduce a new model called "Robust
PCA on Graphs" which incorporates spectral graph regularization into the Robust
PCA framework. Our proposed model benefits from 1) the robustness of principal
components to occlusions and missing values, 2) enhanced low-rank recovery, 3)
improved clustering property due to the graph smoothness assumption on the
low-rank matrix, and 4) convexity of the resulting optimization problem.
Extensive experiments on 8 benchmark, 3 video and 2 artificial datasets with
corruptions clearly reveal that our model outperforms 10 other state-of-the-art
models in its clustering and low-rank recovery tasks
HyperNTF: A Hypergraph Regularized Nonnegative Tensor Factorization for Dimensionality Reduction
Most methods for dimensionality reduction are based on either tensor
representation or local geometry learning. However, the tensor-based methods
severely rely on the assumption of global and multilinear structures in
high-dimensional data; and the manifold learning methods suffer from the
out-of-sample problem. In this paper, bridging the tensor decomposition and
manifold learning, we propose a novel method, called Hypergraph Regularized
Nonnegative Tensor Factorization (HyperNTF). HyperNTF can preserve
nonnegativity in tensor factorization, and uncover the higher-order
relationship among the nearest neighborhoods. Clustering analysis with HyperNTF
has low computation and storage costs. The experiments on four synthetic data
show a desirable property of hypergraph in uncovering the high-order
correlation to unfold the curved manifolds. Moreover, the numerical experiments
on six real datasets suggest that HyperNTF robustly outperforms
state-of-the-art algorithms in clustering analysis.Comment: 12 pages, 6 figures, 9 table
Quadratically-Regularized Optimal Transport on Graphs
Optimal transportation provides a means of lifting distances between points
on a geometric domain to distances between signals over the domain, expressed
as probability distributions. On a graph, transportation problems can be used
to express challenging tasks involving matching supply to demand with minimal
shipment expense; in discrete language, these become minimum-cost network flow
problems. Regularization typically is needed to ensure uniqueness for the
linear ground distance case and to improve optimization convergence;
state-of-the-art techniques employ entropic regularization on the
transportation matrix. In this paper, we explore a quadratic alternative to
entropic regularization for transport over a graph. We theoretically analyze
the behavior of quadratically-regularized graph transport, characterizing how
regularization affects the structure of flows in the regime of small but
nonzero regularization. We further exploit elegant second-order structure in
the dual of this problem to derive an easily-implemented Newton-type
optimization algorithm.Comment: 27 page
- …