8,956 research outputs found
Post-Reconstruction Deconvolution of PET Images by Total Generalized Variation Regularization
Improving the quality of positron emission tomography (PET) images, affected
by low resolution and high level of noise, is a challenging task in nuclear
medicine and radiotherapy. This work proposes a restoration method, achieved
after tomographic reconstruction of the images and targeting clinical
situations where raw data are often not accessible. Based on inverse problem
methods, our contribution introduces the recently developed total generalized
variation (TGV) norm to regularize PET image deconvolution. Moreover, we
stabilize this procedure with additional image constraints such as positivity
and photometry invariance. A criterion for updating and adjusting automatically
the regularization parameter in case of Poisson noise is also presented.
Experiments are conducted on both synthetic data and real patient images.Comment: First published in the Proceedings of the 23rd European Signal
Processing Conference (EUSIPCO-2015) in 2015, published by EURASI
Solving Inverse Problems with Piecewise Linear Estimators: From Gaussian Mixture Models to Structured Sparsity
A general framework for solving image inverse problems is introduced in this
paper. The approach is based on Gaussian mixture models, estimated via a
computationally efficient MAP-EM algorithm. A dual mathematical interpretation
of the proposed framework with structured sparse estimation is described, which
shows that the resulting piecewise linear estimate stabilizes the estimation
when compared to traditional sparse inverse problem techniques. This
interpretation also suggests an effective dictionary motivated initialization
for the MAP-EM algorithm. We demonstrate that in a number of image inverse
problems, including inpainting, zooming, and deblurring, the same algorithm
produces either equal, often significantly better, or very small margin worse
results than the best published ones, at a lower computational cost.Comment: 30 page
Fundamental Limitations of Pixel Based Image Deconvolution in Radio Astronomy
Deconvolution is essential for radio interferometric imaging to produce
scientific quality data because of finite sampling in the Fourier plane. Most
deconvolution algorithms are based on CLEAN which uses a grid of image pixels,
or clean components. A critical matter in this process is the selection of
pixel size for optimal results in deconvolution. As a rule of thumb, the pixel
size is chosen smaller than the resolution dictated by the interferometer. For
images consisting of unresolved (or point like) sources, this approach yields
optimal results. However, for sources that are not point like, in particular
for partially resolved sources, the selection of right pixel size is still an
open issue. In this paper, we investigate the limitations of pixelization in
deconvolving extended sources. In particular, we pursue the usage of
orthonormal basis functions to model extended sources yielding better results
than by using clean components.Comment: 4 pages, 5 figures, the 6th IEEE Sensor Array and Multichannel Signal
Processing worksho
Linear inverse problems with noise: primal and primal-dual splitting
In this paper, we propose two algorithms for solving linear inverse problems
when the observations are corrupted by noise. A proper data fidelity term
(log-likelihood) is introduced to reflect the statistics of the noise (e.g.
Gaussian, Poisson). On the other hand, as a prior, the images to restore are
assumed to be positive and sparsely represented in a dictionary of waveforms.
Piecing together the data fidelity and the prior terms, the solution to the
inverse problem is cast as the minimization of a non-smooth convex functional.
We establish the well-posedness of the optimization problem, characterize the
corresponding minimizers, and solve it by means of primal and primal-dual
proximal splitting algorithms originating from the field of non-smooth convex
optimization theory. Experimental results on deconvolution, inpainting and
denoising with some comparison to prior methods are also reported
Deconvolution under Poisson noise using exact data fidelity and synthesis or analysis sparsity priors
In this paper, we propose a Bayesian MAP estimator for solving the
deconvolution problems when the observations are corrupted by Poisson noise.
Towards this goal, a proper data fidelity term (log-likelihood) is introduced
to reflect the Poisson statistics of the noise. On the other hand, as a prior,
the images to restore are assumed to be positive and sparsely represented in a
dictionary of waveforms such as wavelets or curvelets. Both analysis and
synthesis-type sparsity priors are considered. Piecing together the data
fidelity and the prior terms, the deconvolution problem boils down to the
minimization of non-smooth convex functionals (for each prior). We establish
the well-posedness of each optimization problem, characterize the corresponding
minimizers, and solve them by means of proximal splitting algorithms
originating from the realm of non-smooth convex optimization theory.
Experimental results are conducted to demonstrate the potential applicability
of the proposed algorithms to astronomical imaging datasets
- …