112,030 research outputs found
Information Theoretic Limits for Standard and One-Bit Compressed Sensing with Graph-Structured Sparsity
In this paper, we analyze the information theoretic lower bound on the
necessary number of samples needed for recovering a sparse signal under
different compressed sensing settings. We focus on the weighted graph model, a
model-based framework proposed by Hegde et al. (2015), for standard compressed
sensing as well as for one-bit compressed sensing. We study both the noisy and
noiseless regimes. Our analysis is general in the sense that it applies to any
algorithm used to recover the signal. We carefully construct restricted
ensembles for different settings and then apply Fano's inequality to establish
the lower bound on the necessary number of samples. Furthermore, we show that
our bound is tight for one-bit compressed sensing, while for standard
compressed sensing, our bound is tight up to a logarithmic factor of the number
of non-zero entries in the signal
Recursive Compressed Sensing
We introduce a recursive algorithm for performing compressed sensing on
streaming data. The approach consists of a) recursive encoding, where we sample
the input stream via overlapping windowing and make use of the previous
measurement in obtaining the next one, and b) recursive decoding, where the
signal estimate from the previous window is utilized in order to achieve faster
convergence in an iterative optimization scheme applied to decode the new one.
To remove estimation bias, a two-step estimation procedure is proposed
comprising support set detection and signal amplitude estimation. Estimation
accuracy is enhanced by a non-linear voting method and averaging estimates over
multiple windows. We analyze the computational complexity and estimation error,
and show that the normalized error variance asymptotically goes to zero for
sublinear sparsity. Our simulation results show speed up of an order of
magnitude over traditional CS, while obtaining significantly lower
reconstruction error under mild conditions on the signal magnitudes and the
noise level.Comment: Submitted to IEEE Transactions on Information Theor
Permutation Meets Parallel Compressed Sensing: How to Relax Restricted Isometry Property for 2D Sparse Signals
Traditional compressed sensing considers sampling a 1D signal. For a
multidimensional signal, if reshaped into a vector, the required size of the
sensing matrix becomes dramatically large, which increases the storage and
computational complexity significantly. To solve this problem, we propose to
reshape the multidimensional signal into a 2D signal and sample the 2D signal
using compressed sensing column by column with the same sensing matrix. It is
referred to as parallel compressed sensing, and it has much lower storage and
computational complexity. For a given reconstruction performance of parallel
compressed sensing, if a so-called acceptable permutation is applied to the 2D
signal, we show that the corresponding sensing matrix has a smaller required
order of restricted isometry property condition, and thus, storage and
computation requirements are further lowered. A zigzag-scan-based permutation,
which is shown to be particularly useful for signals satisfying a layer model,
is introduced and investigated. As an application of the parallel compressed
sensing with the zigzag-scan-based permutation, a video compression scheme is
presented. It is shown that the zigzag-scan-based permutation increases the
peak signal-to-noise ratio of reconstructed images and video frames.Comment: 30 pages, 10 figures, 3 tables, submitted to the IEEE Trans. Signal
Processing in November 201
Universal Compressed Sensing
In this paper, the problem of developing universal algorithms for compressed
sensing of stochastic processes is studied. First, R\'enyi's notion of
information dimension (ID) is generalized to analog stationary processes. This
provides a measure of complexity for such processes and is connected to the
number of measurements required for their accurate recovery. Then a minimum
entropy pursuit (MEP) optimization approach is proposed, and it is proven that
it can reliably recover any stationary process satisfying some mixing
constraints from sufficient number of randomized linear measurements, without
having any prior information about the distribution of the process. It is
proved that a Lagrangian-type approximation of the MEP optimization problem,
referred to as Lagrangian-MEP problem, is identical to a heuristic
implementable algorithm proposed by Baron et al. It is shown that for the right
choice of parameters the Lagrangian-MEP algorithm, in addition to having the
same asymptotic performance as MEP optimization, is also robust to the
measurement noise. For memoryless sources with a discrete-continuous mixture
distribution, the fundamental limits of the minimum number of required
measurements by a non-universal compressed sensing decoder is characterized by
Wu et al. For such sources, it is proved that there is no loss in universal
coding, and both the MEP and the Lagrangian-MEP asymptotically achieve the
optimal performance
"Compressed" Compressed Sensing
The field of compressed sensing has shown that a sparse but otherwise
arbitrary vector can be recovered exactly from a small number of randomly
constructed linear projections (or samples). The question addressed in this
paper is whether an even smaller number of samples is sufficient when there
exists prior knowledge about the distribution of the unknown vector, or when
only partial recovery is needed. An information-theoretic lower bound with
connections to free probability theory and an upper bound corresponding to a
computationally simple thresholding estimator are derived. It is shown that in
certain cases (e.g. discrete valued vectors or large distortions) the number of
samples can be decreased. Interestingly though, it is also shown that in many
cases no reduction is possible
- …
