974 research outputs found

    Noisy Signal Recovery via Iterative Reweighted L1-Minimization

    Get PDF
    Compressed sensing has shown that it is possible to reconstruct sparse high dimensional signals from few linear measurements. In many cases, the solution can be obtained by solving an L1-minimization problem, and this method is accurate even in the presence of noise. Recent a modified version of this method, reweighted L1-minimization, has been suggested. Although no provable results have yet been attained, empirical studies have suggested the reweighted version outperforms the standard method. Here we analyze the reweighted L1-minimization method in the noisy case, and provide provable results showing an improvement in the error bound over the standard bounds

    On the Mathematics of Music: From Chords to Fourier Analysis

    Full text link
    Mathematics is a far reaching discipline and its tools appear in many applications. In this paper we discuss its role in music and signal processing by revisiting the use of mathematics in algorithms that can extract chord information from recorded music. We begin with a light introduction to the theory of music and motivate the use of Fourier analysis in audio processing. We introduce the discrete and continuous Fourier transforms and investigate their use in extracting important information from audio data

    Unicity conditions for low-rank matrix recovery

    Get PDF
    Low-rank matrix recovery addresses the problem of recovering an unknown low-rank matrix from few linear measurements. Nuclear-norm minimization is a tractible approach with a recent surge of strong theoretical backing. Analagous to the theory of compressed sensing, these results have required random measurements. For example, m >= Cnr Gaussian measurements are sufficient to recover any rank-r n x n matrix with high probability. In this paper we address the theoretical question of how many measurements are needed via any method whatsoever --- tractible or not. We show that for a family of random measurement ensembles, m >= 4nr - 4r^2 measurements are sufficient to guarantee that no rank-2r matrix lies in the null space of the measurement operator with probability one. This is a necessary and sufficient condition to ensure uniform recovery of all rank-r matrices by rank minimization. Furthermore, this value of mm precisely matches the dimension of the manifold of all rank-2r matrices. We also prove that for a fixed rank-r matrix, m >= 2nr - r^2 + 1 random measurements are enough to guarantee recovery using rank minimization. These results give a benchmark to which we may compare the efficacy of nuclear-norm minimization

    An Asynchronous Parallel Approach to Sparse Recovery

    Full text link
    Asynchronous parallel computing and sparse recovery are two areas that have received recent interest. Asynchronous algorithms are often studied to solve optimization problems where the cost function takes the form βˆ‘i=1Mfi(x)\sum_{i=1}^M f_i(x), with a common assumption that each fif_i is sparse; that is, each fif_i acts only on a small number of components of x∈Rnx\in\mathbb{R}^n. Sparse recovery problems, such as compressed sensing, can be formulated as optimization problems, however, the cost functions fif_i are dense with respect to the components of xx, and instead the signal xx is assumed to be sparse, meaning that it has only ss non-zeros where sβ‰ͺns\ll n. Here we address how one may use an asynchronous parallel architecture when the cost functions fif_i are not sparse in xx, but rather the signal xx is sparse. We propose an asynchronous parallel approach to sparse recovery via a stochastic greedy algorithm, where multiple processors asynchronously update a vector in shared memory containing information on the estimated signal support. We include numerical simulations that illustrate the potential benefits of our proposed asynchronous method.Comment: 5 pages, 2 figure

    Two-subspace Projection Method for Coherent Overdetermined Systems

    Get PDF
    We present a Projection onto Convex Sets (POCS) type algorithm for solving systems of linear equations. POCS methods have found many applications ranging from computer tomography to digital signal and image processing. The Kaczmarz method is one of the most popular solvers for overdetermined systems of linear equations due to its speed and simplicity. Here we introduce and analyze an extension of the Kaczmarz method that iteratively projects the estimate onto a solution space given by two randomly selected rows. We show that this projection algorithm provides exponential convergence to the solution in expectation. The convergence rate improves upon that of the standard randomized Kaczmarz method when the system has correlated rows. Experimental results confirm that in this case our method significantly outperforms the randomized Kaczmarz method.Comment: arXiv admin note: substantial text overlap with arXiv:1204.027

    Stable image reconstruction using total variation minimization

    Get PDF
    This article presents near-optimal guarantees for accurate and robust image recovery from under-sampled noisy measurements using total variation minimization. In particular, we show that from O(slog(N)) nonadaptive linear measurements, an image can be reconstructed to within the best s-term approximation of its gradient up to a logarithmic factor, and this factor can be removed by taking slightly more measurements. Along the way, we prove a strengthened Sobolev inequality for functions lying in the null space of suitably incoherent matrices.Comment: 25 page
    • …
    corecore