5,952 research outputs found
Compressive Signal Processing with Circulant Sensing Matrices
Compressive sensing achieves effective dimensionality reduction of signals,
under a sparsity constraint, by means of a small number of random measurements
acquired through a sensing matrix. In a signal processing system, the problem
arises of processing the random projections directly, without first
reconstructing the signal. In this paper, we show that circulant sensing
matrices allow to perform a variety of classical signal processing tasks such
as filtering, interpolation, registration, transforms, and so forth, directly
in the compressed domain and in an exact fashion, \emph{i.e.}, without relying
on estimators as proposed in the existing literature. The advantage of the
techniques presented in this paper is to enable direct
measurement-to-measurement transformations, without the need of costly recovery
procedures
Robust one-bit compressed sensing with partial circulant matrices
We present optimal sample complexity estimates for one-bit compressed sensing
problems in a realistic scenario: the procedure uses a structured matrix (a
randomly sub-sampled circulant matrix) and is robust to analog pre-quantization
noise as well as to adversarial bit corruptions in the quantization process.
Our results imply that quantization is not a statistically expensive procedure
in the presence of nontrivial analog noise: recovery requires the same sample
size one would have needed had the measurement matrix been Gaussian and the
noisy analog measurements been given as data
On the Phase Transition of Corrupted Sensing
In \cite{FOY2014}, a sharp phase transition has been numerically observed
when a constrained convex procedure is used to solve the corrupted sensing
problem. In this paper, we present a theoretical analysis for this phenomenon.
Specifically, we establish the threshold below which this convex procedure
fails to recover signal and corruption with high probability. Together with the
work in \cite{FOY2014}, we prove that a sharp phase transition occurs around
the sum of the squares of spherical Gaussian widths of two tangent cones.
Numerical experiments are provided to demonstrate the correctness and sharpness
of our results.Comment: To appear in Proceedings of IEEE International Symposium on
Information Theory 201
Simple Bounds for Noisy Linear Inverse Problems with Exact Side Information
This paper considers the linear inverse problem where we wish to estimate a
structured signal from its corrupted observations. When the problem is
ill-posed, it is natural to make use of a convex function that
exploits the structure of the signal. For example, norm can be used
for sparse signals. To carry out the estimation, we consider two well-known
convex programs: 1) Second order cone program (SOCP), and, 2) Lasso. Assuming
Gaussian measurements, we show that, if precise information about the value
or the -norm of the noise is available, one can do a
particularly good job at estimation. In particular, the reconstruction error
becomes proportional to the "sparsity" of the signal rather than the ambient
dimension of the noise vector. We connect our results to existing works and
provide a discussion on the relation of our results to the standard
least-squares problem. Our error bounds are non-asymptotic and sharp, they
apply to arbitrary convex functions and do not assume any distribution on the
noise.Comment: 13 page
- …