7 research outputs found

    Sparse MRI for motion correction

    Full text link
    MR image sparsity/compressibility has been widely exploited for imaging acceleration with the development of compressed sensing. A sparsity-based approach to rigid-body motion correction is presented for the first time in this paper. A motion is sought after such that the compensated MR image is maximally sparse/compressible among the infinite candidates. Iterative algorithms are proposed that jointly estimate the motion and the image content. The proposed method has a lot of merits, such as no need of additional data and loose requirement for the sampling sequence. Promising results are presented to demonstrate its performance.Comment: To appear in Proceedings of ISBI 2013. 4 pages, 1 figur

    Phase Retrieval for Sparse Signals

    Full text link
    The aim of this paper is to build up the theoretical framework for the recovery of sparse signals from the magnitude of the measurement. We first investigate the minimal number of measurements for the success of the recovery of sparse signals without the phase information. We completely settle the minimality question for the real case and give a lower bound for the complex case. We then study the recovery performance of the β„“1\ell_1 minimization. In particular, we present the null space property which, to our knowledge, is the first sufficient and necessary condition for the success of β„“1\ell_1 minimization for kk-sparse phase retrievable.Comment: 14 page

    Hadamard Wirtinger Flow for Sparse Phase Retrieval

    Full text link
    We consider the problem of reconstructing an nn-dimensional kk-sparse signal from a set of noiseless magnitude-only measurements. Formulating the problem as an unregularized empirical risk minimization task, we study the sample complexity performance of gradient descent with Hadamard parametrization, which we call Hadamard Wirtinger flow (HWF). Provided knowledge of the signal sparsity kk, we prove that a single step of HWF is able to recover the support from k(xmaxβˆ—)βˆ’2k(x^*_{max})^{-2} (modulo logarithmic term) samples, where xmaxβˆ—x^*_{max} is the largest component of the signal in magnitude. This support recovery procedure can be used to initialize existing reconstruction methods and yields algorithms with total runtime proportional to the cost of reading the data and improved sample complexity, which is linear in kk when the signal contains at least one large component. We numerically investigate the performance of HWF at convergence and show that, while not requiring any explicit form of regularization nor knowledge of kk, HWF adapts to the signal sparsity and reconstructs sparse signals with fewer measurements than existing gradient based methods
    corecore