5,033 research outputs found
Empirical Bayes selection of wavelet thresholds
This paper explores a class of empirical Bayes methods for level-dependent
threshold selection in wavelet shrinkage. The prior considered for each wavelet
coefficient is a mixture of an atom of probability at zero and a heavy-tailed
density. The mixing weight, or sparsity parameter, for each level of the
transform is chosen by marginal maximum likelihood. If estimation is carried
out using the posterior median, this is a random thresholding procedure; the
estimation can also be carried out using other thresholding rules with the same
threshold. Details of the calculations needed for implementing the procedure
are included. In practice, the estimates are quick to compute and there is
software available. Simulations on the standard model functions show excellent
performance, and applications to data drawn from various fields of application
are used to explore the practical performance of the approach. By using a
general result on the risk of the corresponding marginal maximum likelihood
approach for a single sequence, overall bounds on the risk of the method are
found subject to membership of the unknown function in one of a wide range of
Besov classes, covering also the case of f of bounded variation. The rates
obtained are optimal for any value of the parameter p in (0,\infty],
simultaneously for a wide range of loss functions, each dominating the L_q norm
of the \sigmath derivative, with \sigma\ge0 and 0<q\le2.Comment: Published at http://dx.doi.org/10.1214/009053605000000345 in the
Annals of Statistics (http://www.imstat.org/aos/) by the Institute of
Mathematical Statistics (http://www.imstat.org
Wavelet Estimators in Nonparametric Regression: A Comparative Simulation Study
Wavelet analysis has been found to be a powerful tool for the nonparametric estimation of spatially-variable objects. We discuss in detail wavelet methods in nonparametric regression, where the data are modelled as observations of a signal contaminated with additive Gaussian noise, and provide an extensive review of the vast literature of wavelet shrinkage and wavelet thresholding estimators developed to denoise such data. These estimators arise from a wide range of classical and empirical Bayes methods treating either individual or blocks of wavelet coefficients. We compare various estimators in an extensive simulation study on a variety of sample sizes, test functions, signal-to-noise ratios and wavelet filters. Because there is no single criterion that can adequately summarise the behaviour of an estimator, we use various criteria to measure performance in finite sample situations. Insight into the performance of these estimators is obtained from graphical outputs and numerical tables. In order to provide some hints of how these estimators should be used to analyse real data sets, a detailed practical step-by-step illustration of a wavelet denoising analysis on electrical consumption is provided. Matlab codes are provided so that all figures and tables in this paper can be reproduced
Sketching for Large-Scale Learning of Mixture Models
Learning parameters from voluminous data can be prohibitive in terms of
memory and computational requirements. We propose a "compressive learning"
framework where we estimate model parameters from a sketch of the training
data. This sketch is a collection of generalized moments of the underlying
probability distribution of the data. It can be computed in a single pass on
the training set, and is easily computable on streams or distributed datasets.
The proposed framework shares similarities with compressive sensing, which aims
at drastically reducing the dimension of high-dimensional signals while
preserving the ability to reconstruct them. To perform the estimation task, we
derive an iterative algorithm analogous to sparse reconstruction algorithms in
the context of linear inverse problems. We exemplify our framework with the
compressive estimation of a Gaussian Mixture Model (GMM), providing heuristics
on the choice of the sketching procedure and theoretical guarantees of
reconstruction. We experimentally show on synthetic data that the proposed
algorithm yields results comparable to the classical Expectation-Maximization
(EM) technique while requiring significantly less memory and fewer computations
when the number of database elements is large. We further demonstrate the
potential of the approach on real large-scale data (over 10 8 training samples)
for the task of model-based speaker verification. Finally, we draw some
connections between the proposed framework and approximate Hilbert space
embedding of probability distributions using random features. We show that the
proposed sketching operator can be seen as an innovative method to design
translation-invariant kernels adapted to the analysis of GMMs. We also use this
theoretical framework to derive information preservation guarantees, in the
spirit of infinite-dimensional compressive sensing
MinMax Radon Barcodes for Medical Image Retrieval
Content-based medical image retrieval can support diagnostic decisions by
clinical experts. Examining similar images may provide clues to the expert to
remove uncertainties in his/her final diagnosis. Beyond conventional feature
descriptors, binary features in different ways have been recently proposed to
encode the image content. A recent proposal is "Radon barcodes" that employ
binarized Radon projections to tag/annotate medical images with content-based
binary vectors, called barcodes. In this paper, MinMax Radon barcodes are
introduced which are superior to "local thresholding" scheme suggested in the
literature. Using IRMA dataset with 14,410 x-ray images from 193 different
classes, the advantage of using MinMax Radon barcodes over \emph{thresholded}
Radon barcodes are demonstrated. The retrieval error for direct search drops by
more than 15\%. As well, SURF, as a well-established non-binary approach, and
BRISK, as a recent binary method are examined to compare their results with
MinMax Radon barcodes when retrieving images from IRMA dataset. The results
demonstrate that MinMax Radon barcodes are faster and more accurate when
applied on IRMA images.Comment: To appear in proceedings of the 12th International Symposium on
Visual Computing, December 12-14, 2016, Las Vegas, Nevada, US
A proximal iteration for deconvolving Poisson noisy images using sparse representations
We propose an image deconvolution algorithm when the data is contaminated by
Poisson noise. The image to restore is assumed to be sparsely represented in a
dictionary of waveforms such as the wavelet or curvelet transforms. Our key
contributions are: First, we handle the Poisson noise properly by using the
Anscombe variance stabilizing transform leading to a {\it non-linear}
degradation equation with additive Gaussian noise. Second, the deconvolution
problem is formulated as the minimization of a convex functional with a
data-fidelity term reflecting the noise properties, and a non-smooth
sparsity-promoting penalties over the image representation coefficients (e.g.
-norm). Third, a fast iterative backward-forward splitting algorithm is
proposed to solve the minimization problem. We derive existence and uniqueness
conditions of the solution, and establish convergence of the iterative
algorithm. Finally, a GCV-based model selection procedure is proposed to
objectively select the regularization parameter. Experimental results are
carried out to show the striking benefits gained from taking into account the
Poisson statistics of the noise. These results also suggest that using
sparse-domain regularization may be tractable in many deconvolution
applications with Poisson noise such as astronomy and microscopy
Improving the performance of translation wavelet transform using BMICA
Research has shown Wavelet Transform to be one of the best methods for denoising biosignals. Translation-Invariant
form of this method has been found to be the best performance. In this paper however we utilize this method and merger with our newly created Independent Component Analysis method â BMICA. Different EEG signals are used to verify the method within the MATLAB environment. Results are then compared with those of the actual Translation-Invariant algorithm and evaluated using the performance measures Mean Square Error (MSE), Peak Signal to Noise Ratio (PSNR), Signal to Distortion Ratio (SDR), and Signal to Interference Ratio (SIR). Experiments revealed that the BMICA Translation-Invariant Wavelet Transform out performed in all four measures. This indicates that it performed superior to the basic Translation- Invariant Wavelet Transform algorithm producing cleaner EEG signals which can influence diagnosis as well as clinical studies of the brain
Image interpolation using Shearlet based iterative refinement
This paper proposes an image interpolation algorithm exploiting sparse
representation for natural images. It involves three main steps: (a) obtaining
an initial estimate of the high resolution image using linear methods like FIR
filtering, (b) promoting sparsity in a selected dictionary through iterative
thresholding, and (c) extracting high frequency information from the
approximation to refine the initial estimate. For the sparse modeling, a
shearlet dictionary is chosen to yield a multiscale directional representation.
The proposed algorithm is compared to several state-of-the-art methods to
assess its objective as well as subjective performance. Compared to the cubic
spline interpolation method, an average PSNR gain of around 0.8 dB is observed
over a dataset of 200 images
ShearLab 3D: Faithful Digital Shearlet Transforms based on Compactly Supported Shearlets
Wavelets and their associated transforms are highly efficient when
approximating and analyzing one-dimensional signals. However, multivariate
signals such as images or videos typically exhibit curvilinear singularities,
which wavelets are provably deficient of sparsely approximating and also of
analyzing in the sense of, for instance, detecting their direction. Shearlets
are a directional representation system extending the wavelet framework, which
overcomes those deficiencies. Similar to wavelets, shearlets allow a faithful
implementation and fast associated transforms. In this paper, we will introduce
a comprehensive carefully documented software package coined ShearLab 3D
(www.ShearLab.org) and discuss its algorithmic details. This package provides
MATLAB code for a novel faithful algorithmic realization of the 2D and 3D
shearlet transform (and their inverses) associated with compactly supported
universal shearlet systems incorporating the option of using CUDA. We will
present extensive numerical experiments in 2D and 3D concerning denoising,
inpainting, and feature extraction, comparing the performance of ShearLab 3D
with similar transform-based algorithms such as curvelets, contourlets, or
surfacelets. In the spirit of reproducible reseaerch, all scripts are
accessible on www.ShearLab.org.Comment: There is another shearlet software package
(http://www.mathematik.uni-kl.de/imagepro/members/haeuser/ffst/) by S.
H\"auser and G. Steidl. We will include this in a revisio
- âŠ