3,405 research outputs found
Projection Pursuit through -Divergence Minimisation
Consider a defined density on a set of very large dimension. It is quite
difficult to find an estimate of this density from a data set. However, it is
possible through a projection pursuit methodology to solve this problem.
Touboul's article "Projection Pursuit Through Relative Entropy Minimization",
2009, demonstrates the interest of the author's method in a very simple given
case. He considers the factorization of a density through an Elliptical
component and some residual density. The above Touboul's work is based on
minimizing relative entropy. In the present article, our proposal will aim at
extending this very methodology to the divergence. Furthermore, we will
also consider the case when the density to be factorized is estimated from an
i.i.d. sample. We will then propose a test for the factorization of the
estimated density. Applications include a new test of fit pertaining to the
Elliptical copulas.Comment: 32 pages, 4 figures, 5 tableaux, elsarticle clas
Compressed sensing imaging techniques for radio interferometry
Radio interferometry probes astrophysical signals through incomplete and
noisy Fourier measurements. The theory of compressed sensing demonstrates that
such measurements may actually suffice for accurate reconstruction of sparse or
compressible signals. We propose new generic imaging techniques based on convex
optimization for global minimization problems defined in this context. The
versatility of the framework notably allows introduction of specific prior
information on the signals, which offers the possibility of significant
improvements of reconstruction relative to the standard local matching pursuit
algorithm CLEAN used in radio astronomy. We illustrate the potential of the
approach by studying reconstruction performances on simulations of two
different kinds of signals observed with very generic interferometric
configurations. The first kind is an intensity field of compact astrophysical
objects. The second kind is the imprint of cosmic strings in the temperature
field of the cosmic microwave background radiation, of particular interest for
cosmology.Comment: 10 pages, 1 figure. Version 2 matches version accepted for
publication in MNRAS. Changes includes: writing corrections, clarifications
of arguments, figure update, and a new subsection 4.1 commenting on the exact
compliance of radio interferometric measurements with compressed sensin
Structured Sparsity: Discrete and Convex approaches
Compressive sensing (CS) exploits sparsity to recover sparse or compressible
signals from dimensionality reducing, non-adaptive sensing mechanisms. Sparsity
is also used to enhance interpretability in machine learning and statistics
applications: While the ambient dimension is vast in modern data analysis
problems, the relevant information therein typically resides in a much lower
dimensional space. However, many solutions proposed nowadays do not leverage
the true underlying structure. Recent results in CS extend the simple sparsity
idea to more sophisticated {\em structured} sparsity models, which describe the
interdependency between the nonzero components of a signal, allowing to
increase the interpretability of the results and lead to better recovery
performance. In order to better understand the impact of structured sparsity,
in this chapter we analyze the connections between the discrete models and
their convex relaxations, highlighting their relative advantages. We start with
the general group sparse model and then elaborate on two important special
cases: the dispersive and the hierarchical models. For each, we present the
models in their discrete nature, discuss how to solve the ensuing discrete
problems and then describe convex relaxations. We also consider more general
structures as defined by set functions and present their convex proxies.
Further, we discuss efficient optimization solutions for structured sparsity
problems and illustrate structured sparsity in action via three applications.Comment: 30 pages, 18 figure
Distributed Representation of Geometrically Correlated Images with Compressed Linear Measurements
This paper addresses the problem of distributed coding of images whose
correlation is driven by the motion of objects or positioning of the vision
sensors. It concentrates on the problem where images are encoded with
compressed linear measurements. We propose a geometry-based correlation model
in order to describe the common information in pairs of images. We assume that
the constitutive components of natural images can be captured by visual
features that undergo local transformations (e.g., translation) in different
images. We first identify prominent visual features by computing a sparse
approximation of a reference image with a dictionary of geometric basis
functions. We then pose a regularized optimization problem to estimate the
corresponding features in correlated images given by quantized linear
measurements. The estimated features have to comply with the compressed
information and to represent consistent transformation between images. The
correlation model is given by the relative geometric transformations between
corresponding features. We then propose an efficient joint decoding algorithm
that estimates the compressed images such that they stay consistent with both
the quantized measurements and the correlation model. Experimental results show
that the proposed algorithm effectively estimates the correlation between
images in multi-view datasets. In addition, the proposed algorithm provides
effective decoding performance that compares advantageously to independent
coding solutions as well as state-of-the-art distributed coding schemes based
on disparity learning
Greed is good: algorithmic results for sparse approximation
This article presents new results on using a greedy algorithm, orthogonal matching pursuit (OMP), to solve the sparse approximation problem over redundant dictionaries. It provides a sufficient condition under which both OMP and Donoho's basis pursuit (BP) paradigm can recover the optimal representation of an exactly sparse signal. It leverages this theory to show that both OMP and BP succeed for every sparse input signal from a wide class of dictionaries. These quasi-incoherent dictionaries offer a natural generalization of incoherent dictionaries, and the cumulative coherence function is introduced to quantify the level of incoherence. This analysis unifies all the recent results on BP and extends them to OMP. Furthermore, the paper develops a sufficient condition under which OMP can identify atoms from an optimal approximation of a nonsparse signal. From there, it argues that OMP is an approximation algorithm for the sparse problem over a quasi-incoherent dictionary. That is, for every input signal, OMP calculates a sparse approximant whose error is only a small factor worse than the minimal error that can be attained with the same number of terms
- âŠ