2,701 research outputs found
Signal Recovery From Random Measurements Via Orthogonal Matching Pursuit
This paper demonstrates theoretically and empirically that a greedy algorithm called Orthogonal Matching Pursuit (OMP) can reliably recover a signal with nonzero entries in dimension given random linear measurements of that signal. This is a massive improvement over previous results, which require measurements. The new results for OMP are comparable with recent results for another approach called Basis Pursuit (BP). In some settings, the OMP algorithm is faster and easier to implement, so it is an attractive alternative to BP for signal recovery problems
Sparse Approximation Via Iterative Thresholding
The well-known shrinkage technique is still relevant for contemporary signal processing problems over redundant dictionaries. We present theoretical and empirical analyses for two iterative algorithms for sparse approximation that use shrinkage. The GENERAL IT algorithm amounts to a Landweber iteration with nonlinear shrinkage at each iteration step. The BLOCK IT algorithm arises in morphological components analysis. A sufficient condition for which General IT exactly recovers a sparse signal is presented, in which the cumulative coherence function naturally arises. This analysis extends previous results concerning the Orthogonal Matching Pursuit (OMP) and Basis Pursuit (BP) algorithms to IT algorithms
List decoding of noisy Reed-Muller-like codes
First- and second-order Reed-Muller (RM(1) and RM(2), respectively) codes are
two fundamental error-correcting codes which arise in communication as well as
in probabilistically-checkable proofs and learning. In this paper, we take the
first steps toward extending the quick randomized decoding tools of RM(1) into
the realm of quadratic binary and, equivalently, Z_4 codes. Our main
algorithmic result is an extension of the RM(1) techniques from Goldreich-Levin
and Kushilevitz-Mansour algorithms to the Hankel code, a code between RM(1) and
RM(2). That is, given signal s of length N, we find a list that is a superset
of all Hankel codewords phi with dot product to s at least (1/sqrt(k)) times
the norm of s, in time polynomial in k and log(N). We also give a new and
simple formulation of a known Kerdock code as a subcode of the Hankel code. As
a corollary, we can list-decode Kerdock, too. Also, we get a quick algorithm
for finding a sparse Kerdock approximation. That is, for k small compared with
1/sqrt{N} and for epsilon > 0, we find, in time polynomial in (k
log(N)/epsilon), a k-Kerdock-term approximation s~ to s with Euclidean error at
most the factor (1+epsilon+O(k^2/sqrt{N})) times that of the best such
approximation
Approximate Sparse Recovery: Optimizing Time and Measurements
An approximate sparse recovery system consists of parameters , an
-by- measurement matrix, , and a decoding algorithm, .
Given a vector, , the system approximates by , which must satisfy , where denotes the optimal -term approximation to . For
each vector , the system must succeed with probability at least 3/4. Among
the goals in designing such systems are minimizing the number of
measurements and the runtime of the decoding algorithm, .
In this paper, we give a system with
measurements--matching a lower bound, up to a constant factor--and decoding
time , matching a lower bound up to factors.
We also consider the encode time (i.e., the time to multiply by ),
the time to update measurements (i.e., the time to multiply by a
1-sparse ), and the robustness and stability of the algorithm (adding noise
before and after the measurements). Our encode and update times are optimal up
to factors
- …