1,836 research outputs found
On the Effective Measure of Dimension in the Analysis Cosparse Model
Many applications have benefited remarkably from low-dimensional models in
the recent decade. The fact that many signals, though high dimensional, are
intrinsically low dimensional has given the possibility to recover them stably
from a relatively small number of their measurements. For example, in
compressed sensing with the standard (synthesis) sparsity prior and in matrix
completion, the number of measurements needed is proportional (up to a
logarithmic factor) to the signal's manifold dimension.
Recently, a new natural low-dimensional signal model has been proposed: the
cosparse analysis prior. In the noiseless case, it is possible to recover
signals from this model, using a combinatorial search, from a number of
measurements proportional to the signal's manifold dimension. However, if we
ask for stability to noise or an efficient (polynomial complexity) solver, all
the existing results demand a number of measurements which is far removed from
the manifold dimension, sometimes far greater. Thus, it is natural to ask
whether this gap is a deficiency of the theory and the solvers, or if there
exists a real barrier in recovering the cosparse signals by relying only on
their manifold dimension. Is there an algorithm which, in the presence of
noise, can accurately recover a cosparse signal from a number of measurements
proportional to the manifold dimension? In this work, we prove that there is no
such algorithm. Further, we show through numerical simulations that even in the
noiseless case convex relaxations fail when the number of measurements is
comparable to the manifold dimension. This gives a practical counter-example to
the growing literature on compressed acquisition of signals based on manifold
dimension.Comment: 19 pages, 6 figure
Sampling and Recovery of Pulse Streams
Compressive Sensing (CS) is a new technique for the efficient acquisition of
signals, images, and other data that have a sparse representation in some
basis, frame, or dictionary. By sparse we mean that the N-dimensional basis
representation has just K<<N significant coefficients; in this case, the CS
theory maintains that just M = K log N random linear signal measurements will
both preserve all of the signal information and enable robust signal
reconstruction in polynomial time. In this paper, we extend the CS theory to
pulse stream data, which correspond to S-sparse signals/images that are
convolved with an unknown F-sparse pulse shape. Ignoring their convolutional
structure, a pulse stream signal is K=SF sparse. Such signals figure
prominently in a number of applications, from neuroscience to astronomy. Our
specific contributions are threefold. First, we propose a pulse stream signal
model and show that it is equivalent to an infinite union of subspaces. Second,
we derive a lower bound on the number of measurements M required to preserve
the essential information present in pulse streams. The bound is linear in the
total number of degrees of freedom S + F, which is significantly smaller than
the naive bound based on the total signal sparsity K=SF. Third, we develop an
efficient signal recovery algorithm that infers both the shape of the impulse
response as well as the locations and amplitudes of the pulses. The algorithm
alternatively estimates the pulse locations and the pulse shape in a manner
reminiscent of classical deconvolution algorithms. Numerical experiments on
synthetic and real data demonstrate the advantages of our approach over
standard CS
Sub-Nyquist Sampling: Bridging Theory and Practice
Sampling theory encompasses all aspects related to the conversion of
continuous-time signals to discrete streams of numbers. The famous
Shannon-Nyquist theorem has become a landmark in the development of digital
signal processing. In modern applications, an increasingly number of functions
is being pushed forward to sophisticated software algorithms, leaving only
those delicate finely-tuned tasks for the circuit level.
In this paper, we review sampling strategies which target reduction of the
ADC rate below Nyquist. Our survey covers classic works from the early 50's of
the previous century through recent publications from the past several years.
The prime focus is bridging theory and practice, that is to pinpoint the
potential of sub-Nyquist strategies to emerge from the math to the hardware. In
that spirit, we integrate contemporary theoretical viewpoints, which study
signal modeling in a union of subspaces, together with a taste of practical
aspects, namely how the avant-garde modalities boil down to concrete signal
processing systems. Our hope is that this presentation style will attract the
interest of both researchers and engineers in the hope of promoting the
sub-Nyquist premise into practical applications, and encouraging further
research into this exciting new frontier.Comment: 48 pages, 18 figures, to appear in IEEE Signal Processing Magazin
Compressive Sensing of Analog Signals Using Discrete Prolate Spheroidal Sequences
Compressive sensing (CS) has recently emerged as a framework for efficiently
capturing signals that are sparse or compressible in an appropriate basis.
While often motivated as an alternative to Nyquist-rate sampling, there remains
a gap between the discrete, finite-dimensional CS framework and the problem of
acquiring a continuous-time signal. In this paper, we attempt to bridge this
gap by exploiting the Discrete Prolate Spheroidal Sequences (DPSS's), a
collection of functions that trace back to the seminal work by Slepian, Landau,
and Pollack on the effects of time-limiting and bandlimiting operations. DPSS's
form a highly efficient basis for sampled bandlimited functions; by modulating
and merging DPSS bases, we obtain a dictionary that offers high-quality sparse
approximations for most sampled multiband signals. This multiband modulated
DPSS dictionary can be readily incorporated into the CS framework. We provide
theoretical guarantees and practical insight into the use of this dictionary
for recovery of sampled multiband signals from compressive measurements
Time Delay Estimation from Low Rate Samples: A Union of Subspaces Approach
Time delay estimation arises in many applications in which a multipath medium
has to be identified from pulses transmitted through the channel. Various
approaches have been proposed in the literature to identify time delays
introduced by multipath environments. However, these methods either operate on
the analog received signal, or require high sampling rates in order to achieve
reasonable time resolution. In this paper, our goal is to develop a unified
approach to time delay estimation from low rate samples of the output of a
multipath channel. Our methods result in perfect recovery of the multipath
delays from samples of the channel output at the lowest possible rate, even in
the presence of overlapping transmitted pulses. This rate depends only on the
number of multipath components and the transmission rate, but not on the
bandwidth of the probing signal. In addition, our development allows for a
variety of different sampling methods. By properly manipulating the low-rate
samples, we show that the time delays can be recovered using the well-known
ESPRIT algorithm. Combining results from sampling theory with those obtained in
the context of direction of arrival estimation methods, we develop necessary
and sufficient conditions on the transmitted pulse and the sampling functions
in order to ensure perfect recovery of the channel parameters at the minimal
possible rate. Our results can be viewed in a broader context, as a sampling
theorem for analog signals defined over an infinite union of subspaces
Compressed Sensing of Analog Signals in Shift-Invariant Spaces
A traditional assumption underlying most data converters is that the signal
should be sampled at a rate exceeding twice the highest frequency. This
statement is based on a worst-case scenario in which the signal occupies the
entire available bandwidth. In practice, many signals are sparse so that only
part of the bandwidth is used. In this paper, we develop methods for low-rate
sampling of continuous-time sparse signals in shift-invariant (SI) spaces,
generated by m kernels with period T. We model sparsity by treating the case in
which only k out of the m generators are active, however, we do not know which
k are chosen. We show how to sample such signals at a rate much lower than m/T,
which is the minimal sampling rate without exploiting sparsity. Our approach
combines ideas from analog sampling in a subspace with a recently developed
block diagram that converts an infinite set of sparse equations to a finite
counterpart. Using these two components we formulate our problem within the
framework of finite compressed sensing (CS) and then rely on algorithms
developed in that context. The distinguishing feature of our results is that in
contrast to standard CS, which treats finite-length vectors, we consider
sampling of analog signals for which no underlying finite-dimensional model
exists. The proposed framework allows to extend much of the recent literature
on CS to the analog domain.Comment: to appear in IEEE Trans. on Signal Processin
A Dimension Reduction Scheme for the Computation of Optimal Unions of Subspaces
Given a set of points \F in a high dimensional space, the problem of finding
a union of subspaces \cup_i V_i\subset \R^N that best explains the data \F
increases dramatically with the dimension of \R^N. In this article, we study a
class of transformations that map the problem into another one in lower
dimension. We use the best model in the low dimensional space to approximate
the best solution in the original high dimensional space. We then estimate the
error produced between this solution and the optimal solution in the high
dimensional space.Comment: 15 pages. Some corrections were added, in particular the title was
changed. It will appear in "Sampling Theory in Signal and Image Processing
- …