7,728 research outputs found
Efficient Compressive Sensing with Deterministic Guarantees Using Expander Graphs
Compressive sensing is an emerging technology which can recover a sparse signal vector of dimension n via a much smaller number of measurements than n. However, the existing compressive sensing methods may still suffer from relatively high recovery complexity, such as O(n^3), or can only work efficiently when the signal is super sparse, sometimes without deterministic performance guarantees. In this paper, we propose a compressive sensing scheme with deterministic performance guarantees using expander-graphs-based measurement matrices and show that the signal recovery can be achieved with complexity O(n) even if the number of nonzero elements k grows linearly with n. We also investigate compressive sensing for approximately sparse signals using this new method. Moreover, explicit constructions of the considered expander graphs exist. Simulation results are given to show the performance and complexity of the new method
One-Bit Compressive Sensing of Dictionary-Sparse Signals
One-bit compressive sensing has extended the scope of sparse recovery by showing that sparse signals can be accurately reconstructed even when their linear measurements are subject to the extreme quantization scenario of binary samples—only the sign of each linear measurement is maintained. Existing results in one-bit compressive sensing rely on the assumption that the signals of interest are sparse in some fixed orthonormal basis. However, in most practical applications, signals are sparse with respect to an overcomplete dictionary, rather than a basis. There has already been a surge of activity to obtain recovery guarantees under such a generalized sparsity model in the classical compressive sensing setting. Here, we extend the one-bit framework to this important model, providing a unified theory of one-bit compressive sensing under dictionary sparsity. Specifically, we analyze several different algorithms—based on convex programming and on hard thresholding—and show that, under natural assumptions on the sensing matrix (satisfied by Gaussian matrices), these algorithms can efficiently recover analysis-dictionary-sparse signals in the one-bit model
Nonuniform Sparse Recovery with Subgaussian Matrices
Compressive sensing predicts that sufficiently sparse vectors can be
recovered from highly incomplete information. Efficient recovery methods such
as -minimization find the sparsest solution to certain systems of
equations. Random matrices have become a popular choice for the measurement
matrix. Indeed, near-optimal uniform recovery results have been shown for such
matrices. In this note we focus on nonuniform recovery using Gaussian random
matrices and -minimization. We provide a condition on the number of
samples in terms of the sparsity and the signal length which guarantees that a
fixed sparse signal can be recovered with a random draw of the matrix using
-minimization. The constant 2 in the condition is optimal, and the
proof is rather short compared to a similar result due to Donoho and Tanner
Sparse recovery in bounded Riesz systems with applications to numerical methods for PDEs
We study sparse recovery with structured random measurement matrices having
independent, identically distributed, and uniformly bounded rows and with a
nontrivial covariance structure. This class of matrices arises from random
sampling of bounded Riesz systems and generalizes random partial Fourier
matrices. Our main result improves the currently available results for the null
space and restricted isometry properties of such random matrices. The main
novelty of our analysis is a new upper bound for the expectation of the
supremum of a Bernoulli process associated with a restricted isometry constant.
We apply our result to prove new performance guarantees for the CORSING method,
a recently introduced numerical approximation technique for partial
differential equations (PDEs) based on compressive sensing
- …