1,275 research outputs found
Skellam shrinkage: Wavelet-based intensity estimation for inhomogeneous Poisson data
The ubiquity of integrating detectors in imaging and other applications
implies that a variety of real-world data are well modeled as Poisson random
variables whose means are in turn proportional to an underlying vector-valued
signal of interest. In this article, we first show how the so-called Skellam
distribution arises from the fact that Haar wavelet and filterbank transform
coefficients corresponding to measurements of this type are distributed as sums
and differences of Poisson counts. We then provide two main theorems on Skellam
shrinkage, one showing the near-optimality of shrinkage in the Bayesian setting
and the other providing for unbiased risk estimation in a frequentist context.
These results serve to yield new estimators in the Haar transform domain,
including an unbiased risk estimate for shrinkage of Haar-Fisz
variance-stabilized data, along with accompanying low-complexity algorithms for
inference. We conclude with a simulation study demonstrating the efficacy of
our Skellam shrinkage estimators both for the standard univariate wavelet test
functions as well as a variety of test images taken from the image processing
literature, confirming that they offer substantial performance improvements
over existing alternatives.Comment: 27 pages, 8 figures, slight formatting changes; submitted for
publicatio
Fast Separable Non-Local Means
We propose a simple and fast algorithm called PatchLift for computing
distances between patches (contiguous block of samples) extracted from a given
one-dimensional signal. PatchLift is based on the observation that the patch
distances can be efficiently computed from a matrix that is derived from the
one-dimensional signal using lifting; importantly, the number of operations
required to compute the patch distances using this approach does not scale with
the patch length. We next demonstrate how PatchLift can be used for patch-based
denoising of images corrupted with Gaussian noise. In particular, we propose a
separable formulation of the classical Non-Local Means (NLM) algorithm that can
be implemented using PatchLift. We demonstrate that the PatchLift-based
implementation of separable NLM is few orders faster than standard NLM, and is
competitive with existing fast implementations of NLM. Moreover, its denoising
performance is shown to be consistently superior to that of NLM and some of its
variants, both in terms of PSNR/SSIM and visual quality
A CURE for noisy magnetic resonance images: Chi-square unbiased risk estimation
In this article we derive an unbiased expression for the expected
mean-squared error associated with continuously differentiable estimators of
the noncentrality parameter of a chi-square random variable. We then consider
the task of denoising squared-magnitude magnetic resonance image data, which
are well modeled as independent noncentral chi-square random variables on two
degrees of freedom. We consider two broad classes of linearly parameterized
shrinkage estimators that can be optimized using our risk estimate, one in the
general context of undecimated filterbank transforms, and another in the
specific case of the unnormalized Haar wavelet transform. The resultant
algorithms are computationally tractable and improve upon state-of-the-art
methods for both simulated and actual magnetic resonance image data.Comment: 30 double-spaced pages, 11 figures; submitted for publicatio
Sobolev spaces with non-Muckenhoupt weights, fractional elliptic operators, and applications
We propose a new variational model in weighted Sobolev spaces with
non-standard weights and applications to image processing. We show that these
weights are, in general, not of Muckenhoupt type and therefore the classical
analysis tools may not apply. For special cases of the weights, the resulting
variational problem is known to be equivalent to the fractional Poisson
problem. The trace space for the weighted Sobolev space is identified to be
embedded in a weighted space. We propose a finite element scheme to solve
the Euler-Lagrange equations, and for the image denoising application we
propose an algorithm to identify the unknown weights. The approach is
illustrated on several test problems and it yields better results when compared
to the existing total variation techniques
Joint Total Variation ESTATICS for Robust Multi-Parameter Mapping
Quantitative magnetic resonance imaging (qMRI) derives tissue-specific
parameters -- such as the apparent transverse relaxation rate R2*, the
longitudinal relaxation rate R1 and the magnetisation transfer saturation --
that can be compared across sites and scanners and carry important information
about the underlying microstructure. The multi-parameter mapping (MPM) protocol
takes advantage of multi-echo acquisitions with variable flip angles to extract
these parameters in a clinically acceptable scan time. In this context,
ESTATICS performs a joint loglinear fit of multiple echo series to extract R2*
and multiple extrapolated intercepts, thereby improving robustness to motion
and decreasing the variance of the estimators. In this paper, we extend this
model in two ways: (1) by introducing a joint total variation (JTV) prior on
the intercepts and decay, and (2) by deriving a nonlinear maximum \emph{a
posteriori} estimate. We evaluated the proposed algorithm by predicting
left-out echoes in a rich single-subject dataset. In this validation, we
outperformed other state-of-the-art methods and additionally showed that the
proposed approach greatly reduces the variance of the estimated maps, without
introducing bias.Comment: 11 pages, 2 figures, 1 table, conference paper, accepted at MICCAI
202
Recent Progress in Image Deblurring
This paper comprehensively reviews the recent development of image
deblurring, including non-blind/blind, spatially invariant/variant deblurring
techniques. Indeed, these techniques share the same objective of inferring a
latent sharp image from one or several corresponding blurry images, while the
blind deblurring techniques are also required to derive an accurate blur
kernel. Considering the critical role of image restoration in modern imaging
systems to provide high-quality images under complex environments such as
motion, undesirable lighting conditions, and imperfect system components, image
deblurring has attracted growing attention in recent years. From the viewpoint
of how to handle the ill-posedness which is a crucial issue in deblurring
tasks, existing methods can be grouped into five categories: Bayesian inference
framework, variational methods, sparse representation-based methods,
homography-based modeling, and region-based methods. In spite of achieving a
certain level of development, image deblurring, especially the blind case, is
limited in its success by complex application conditions which make the blur
kernel hard to obtain and be spatially variant. We provide a holistic
understanding and deep insight into image deblurring in this review. An
analysis of the empirical evidence for representative methods, practical
issues, as well as a discussion of promising future directions are also
presented.Comment: 53 pages, 17 figure
A Tutorial on Speckle Reduction in Synthetic Aperture Radar Images
Speckle is a granular disturbance, usually modeled as a multiplicative noise, that affects synthetic aperture radar (SAR) images, as well as all coherent images. Over the last three decades, several methods have been proposed for the reduction of speckle, or despeckling, in SAR images. Goal of this paper is making a comprehensive review of despeckling methods since their birth, over thirty years ago, highlighting trends and changing approaches over years. The concept of fully developed speckle is explained. Drawbacks of homomorphic filtering are pointed out. Assets of multiresolution despeckling, as opposite to spatial-domain despeckling, are highlighted. Also advantages of undecimated, or stationary, wavelet transforms over decimated ones are discussed. Bayesian estimators and probability density function (pdf) models in both spatial and multiresolution domains are reviewed. Scale-space varying pdf models, as opposite to scale varying models, are promoted. Promising methods following non-Bayesian approaches, like nonlocal (NL) filtering and total variation (TV) regularization, are reviewed and compared to spatial- and wavelet-domain Bayesian filters. Both established and new trends for assessment of despeckling are presented. A few experiments on simulated data and real COSMO-SkyMed SAR images highlight, on one side the costperformance tradeoff of the different methods, on the other side the effectiveness of solutions purposely designed for SAR heterogeneity and not fully developed speckle. Eventually, upcoming methods based on new concepts of signal processing, like compressive sensing, are foreseen as a new generation of despeckling, after spatial-domain and multiresolution-domain method
- …