805 research outputs found
Automatic Kalman-filter-based wavelet shrinkage denoising of 1D stellar spectra
We propose a non-parametric method to denoise 1D stellar spectra based on wavelet shrinkage followed by adaptive Kalman thresholding. Wavelet shrinkage denoising involves applying the discrete wavelet transform (DWT) to the input signal, 'shrinking' certain frequency components in the transform domain, and then applying inverse DWT to the reduced components. The performance of this procedure is influenced by the choice of base wavelet, the number of decomposition levels, and the thresholding function. Typically, these parameters are chosen by 'trial and error', which can be strongly dependent on the properties of the data being denoised. We here introduce an adaptive Kalman-filter-based thresholding method that eliminates the need for choosing the number of decomposition levels. We use the 'Haar' wavelet basis, which we found to provide excellent filtering for 1D stellar spectra, at a low computational cost. We introduce various levels of Poisson noise into synthetic PHOENIX spectra, and test the performance of several common denoising methods against our own. It proves superior in terms of noise suppression and peak shape preservation. We expect it may also be of use in automatically and accurately filtering low signal-to-noise galaxy and quasar spectra obtained from surveys such as SDSS, Gaia, LSST, PESSTO, VANDELS, LEGA-C, and DESI
Source detection using a 3D sparse representation: application to the Fermi gamma-ray space telescope
The multiscale variance stabilization Transform (MSVST) has recently been
proposed for Poisson data denoising. This procedure, which is nonparametric, is
based on thresholding wavelet coefficients. We present in this paper an
extension of the MSVST to 3D data (in fact 2D-1D data) when the third dimension
is not a spatial dimension, but the wavelength, the energy, or the time. We
show that the MSVST can be used for detecting and characterizing astrophysical
sources of high-energy gamma rays, using realistic simulated observations with
the Large Area Telescope (LAT). The LAT was launched in June 2008 on the Fermi
Gamma-ray Space Telescope mission. The MSVST algorithm is very fast relative to
traditional likelihood model fitting, and permits efficient detection across
the time dimension and immediate estimation of spectral properties.
Astrophysical sources of gamma rays, especially active galaxies, are typically
quite variable, and our current work may lead to a reliable method to quickly
characterize the flaring properties of newly-detected sources.Comment: Accepted. Full paper will figures available at
http://jstarck.free.fr/aa08_msvst.pd
A Panorama on Multiscale Geometric Representations, Intertwining Spatial, Directional and Frequency Selectivity
The richness of natural images makes the quest for optimal representations in
image processing and computer vision challenging. The latter observation has
not prevented the design of image representations, which trade off between
efficiency and complexity, while achieving accurate rendering of smooth regions
as well as reproducing faithful contours and textures. The most recent ones,
proposed in the past decade, share an hybrid heritage highlighting the
multiscale and oriented nature of edges and patterns in images. This paper
presents a panorama of the aforementioned literature on decompositions in
multiscale, multi-orientation bases or dictionaries. They typically exhibit
redundancy to improve sparsity in the transformed domain and sometimes its
invariance with respect to simple geometric deformations (translation,
rotation). Oriented multiscale dictionaries extend traditional wavelet
processing and may offer rotation invariance. Highly redundant dictionaries
require specific algorithms to simplify the search for an efficient (sparse)
representation. We also discuss the extension of multiscale geometric
decompositions to non-Euclidean domains such as the sphere or arbitrary meshed
surfaces. The etymology of panorama suggests an overview, based on a choice of
partially overlapping "pictures". We hope that this paper will contribute to
the appreciation and apprehension of a stream of current research directions in
image understanding.Comment: 65 pages, 33 figures, 303 reference
MDL Denoising Revisited
We refine and extend an earlier MDL denoising criterion for wavelet-based
denoising. We start by showing that the denoising problem can be reformulated
as a clustering problem, where the goal is to obtain separate clusters for
informative and non-informative wavelet coefficients, respectively. This
suggests two refinements, adding a code-length for the model index, and
extending the model in order to account for subband-dependent coefficient
distributions. A third refinement is derivation of soft thresholding inspired
by predictive universal coding with weighted mixtures. We propose a practical
method incorporating all three refinements, which is shown to achieve good
performance and robustness in denoising both artificial and natural signals.Comment: Submitted to IEEE Transactions on Information Theory, June 200
Evidence for a Galactic gamma ray halo
We present quantitative statistical evidence for a -ray emission halo
surrounding the Galaxy. Maps of the emission are derived. EGRET data were
analyzed in a wavelet-based non-parametric hypothesis testing framework, using
a model of expected diffuse (Galactic + isotropic) emission as a null
hypothesis. The results show a statistically significant large scale halo
surrounding the center of the Milky Way as seen from Earth. The halo flux at
high latitudes is somewhat smaller than the isotropic gamma-ray flux at the
same energy, though of the same order (O(10^(-7)--10^(-6)) ph/cm^2/s/sr above 1
GeV).Comment: Final version accepted for publication in New Astronomy. Some
additional results/discussion included, along with entirely revised figures.
19 pages, 15 figures, AASTeX. Better quality figs (PS and JPEG) are available
at http://tigre.ucr.edu/halo/paper.htm
Constrained Overcomplete Analysis Operator Learning for Cosparse Signal Modelling
We consider the problem of learning a low-dimensional signal model from a
collection of training samples. The mainstream approach would be to learn an
overcomplete dictionary to provide good approximations of the training samples
using sparse synthesis coefficients. This famous sparse model has a less well
known counterpart, in analysis form, called the cosparse analysis model. In
this new model, signals are characterised by their parsimony in a transformed
domain using an overcomplete (linear) analysis operator. We propose to learn an
analysis operator from a training corpus using a constrained optimisation
framework based on L1 optimisation. The reason for introducing a constraint in
the optimisation framework is to exclude trivial solutions. Although there is
no final answer here for which constraint is the most relevant constraint, we
investigate some conventional constraints in the model adaptation field and use
the uniformly normalised tight frame (UNTF) for this purpose. We then derive a
practical learning algorithm, based on projected subgradients and
Douglas-Rachford splitting technique, and demonstrate its ability to robustly
recover a ground truth analysis operator, when provided with a clean training
set, of sufficient size. We also find an analysis operator for images, using
some noisy cosparse signals, which is indeed a more realistic experiment. As
the derived optimisation problem is not a convex program, we often find a local
minimum using such variational methods. Some local optimality conditions are
derived for two different settings, providing preliminary theoretical support
for the well-posedness of the learning problem under appropriate conditions.Comment: 29 pages, 13 figures, accepted to be published in TS
Wavelet Analysis and Denoising: New Tools for Economists
This paper surveys the techniques of wavelets analysis and the associated methods of denoising. The Discrete Wavelet Transform and its undecimated version, the Maximum Overlapping Discrete Wavelet Transform, are described. The methods of wavelets analysis can be used to show how the frequency content of the data varies with time. This allows us to pinpoint in time such events as major structural breaks. The sparse nature of the wavelets representation also facilitates the process of noise reduction by nonlinear wavelet shrinkage , which can be used to reveal the underlying trends in economic data. An application of these techniques to the UK real GDP (1873-2001) is described. The purpose of the analysis is to reveal the true structure of the data - including its local irregularities and abrupt changes - and the results are surprising.Wavelets, Denoising, Structural breaks, Trend estimation
- …