937 research outputs found
Regularized Fourier ptychography using an online plug-and-play algorithm
The plug-and-play priors (PnP) framework has been recently shown to achieve state-of-the-art results in regularized image reconstruction by leveraging a sophisticated denoiser within an iterative algorithm. In this paper, we propose a new online PnP algorithm for Fourier ptychographic microscopy (FPM) based on the accelerated proximal gradient method (APGM). Specifically, the proposed algorithm uses only a subset of measurements, which makes it scalable to a large set of measurements. We validate the algorithm by showing that it can lead to significant performance gains on both simulated and experimental data.https://arxiv.org/abs/1811.00120Published versio
Scene-adapted plug-and-play algorithm with convergence guarantees
Recent frameworks, such as the so-called plug-and-play, allow us to leverage
the developments in image denoising to tackle other, and more involved,
problems in image processing. As the name suggests, state-of-the-art denoisers
are plugged into an iterative algorithm that alternates between a denoising
step and the inversion of the observation operator. While these tools offer
flexibility, the convergence of the resulting algorithm may be difficult to
analyse. In this paper, we plug a state-of-the-art denoiser, based on a
Gaussian mixture model, in the iterations of an alternating direction method of
multipliers and prove the algorithm is guaranteed to converge. Moreover, we
build upon the concept of scene-adapted priors where we learn a model targeted
to a specific scene being imaged, and apply the proposed method to address the
hyperspectral sharpening problem
Regularized Fourier ptychography using an online plug-and-play algorithm
The plug-and-play priors (PnP) framework has been recently shown to achieve state-of-the-art results in regularized image reconstruction by leveraging a sophisticated denoiser within an iterative algorithm. In this paper, we propose a new online PnP algorithm for Fourier ptychographic microscopy (FPM) based on the accelerated proximal gradient method (APGM). Specifically, the proposed algorithm uses only a subset of measurements, which makes it scalable to a large set of measurements. We validate the algorithm by showing that it can lead to significant performance gains on both simulated and experimental data.https://arxiv.org/abs/1811.00120Published versio
Plug-and-Play Methods Provably Converge with Properly Trained Denoisers
Plug-and-play (PnP) is a non-convex framework that integrates modern
denoising priors, such as BM3D or deep learning-based denoisers, into ADMM or
other proximal algorithms. An advantage of PnP is that one can use pre-trained
denoisers when there is not sufficient data for end-to-end training. Although
PnP has been recently studied extensively with great empirical success,
theoretical analysis addressing even the most basic question of convergence has
been insufficient. In this paper, we theoretically establish convergence of
PnP-FBS and PnP-ADMM, without using diminishing stepsizes, under a certain
Lipschitz condition on the denoisers. We then propose real spectral
normalization, a technique for training deep learning-based denoisers to
satisfy the proposed Lipschitz condition. Finally, we present experimental
results validating the theory.Comment: Published in the International Conference on Machine Learning, 201
Learning Deep CNN Denoiser Prior for Image Restoration
Model-based optimization methods and discriminative learning methods have
been the two dominant strategies for solving various inverse problems in
low-level vision. Typically, those two kinds of methods have their respective
merits and drawbacks, e.g., model-based optimization methods are flexible for
handling different inverse problems but are usually time-consuming with
sophisticated priors for the purpose of good performance; in the meanwhile,
discriminative learning methods have fast testing speed but their application
range is greatly restricted by the specialized task. Recent works have revealed
that, with the aid of variable splitting techniques, denoiser prior can be
plugged in as a modular part of model-based optimization methods to solve other
inverse problems (e.g., deblurring). Such an integration induces considerable
advantage when the denoiser is obtained via discriminative learning. However,
the study of integration with fast discriminative denoiser prior is still
lacking. To this end, this paper aims to train a set of fast and effective CNN
(convolutional neural network) denoisers and integrate them into model-based
optimization method to solve other inverse problems. Experimental results
demonstrate that the learned set of denoisers not only achieve promising
Gaussian denoising results but also can be used as prior to deliver good
performance for various low-level vision applications.Comment: Accepted to CVPR 2017. Code: https://github.com/cszn/ircn
Deconvolution under Poisson noise using exact data fidelity and synthesis or analysis sparsity priors
In this paper, we propose a Bayesian MAP estimator for solving the
deconvolution problems when the observations are corrupted by Poisson noise.
Towards this goal, a proper data fidelity term (log-likelihood) is introduced
to reflect the Poisson statistics of the noise. On the other hand, as a prior,
the images to restore are assumed to be positive and sparsely represented in a
dictionary of waveforms such as wavelets or curvelets. Both analysis and
synthesis-type sparsity priors are considered. Piecing together the data
fidelity and the prior terms, the deconvolution problem boils down to the
minimization of non-smooth convex functionals (for each prior). We establish
the well-posedness of each optimization problem, characterize the corresponding
minimizers, and solve them by means of proximal splitting algorithms
originating from the realm of non-smooth convex optimization theory.
Experimental results are conducted to demonstrate the potential applicability
of the proposed algorithms to astronomical imaging datasets
- âŠ