3,980 research outputs found
Convolutional Deblurring for Natural Imaging
In this paper, we propose a novel design of image deblurring in the form of
one-shot convolution filtering that can directly convolve with naturally
blurred images for restoration. The problem of optical blurring is a common
disadvantage to many imaging applications that suffer from optical
imperfections. Despite numerous deconvolution methods that blindly estimate
blurring in either inclusive or exclusive forms, they are practically
challenging due to high computational cost and low image reconstruction
quality. Both conditions of high accuracy and high speed are prerequisites for
high-throughput imaging platforms in digital archiving. In such platforms,
deblurring is required after image acquisition before being stored, previewed,
or processed for high-level interpretation. Therefore, on-the-fly correction of
such images is important to avoid possible time delays, mitigate computational
expenses, and increase image perception quality. We bridge this gap by
synthesizing a deconvolution kernel as a linear combination of Finite Impulse
Response (FIR) even-derivative filters that can be directly convolved with
blurry input images to boost the frequency fall-off of the Point Spread
Function (PSF) associated with the optical blur. We employ a Gaussian low-pass
filter to decouple the image denoising problem for image edge deblurring.
Furthermore, we propose a blind approach to estimate the PSF statistics for two
Gaussian and Laplacian models that are common in many imaging pipelines.
Thorough experiments are designed to test and validate the efficiency of the
proposed method using 2054 naturally blurred images across six imaging
applications and seven state-of-the-art deconvolution methods.Comment: 15 pages, for publication in IEEE Transaction Image Processin
FASTLens (FAst STatistics for weak Lensing) : Fast method for Weak Lensing Statistics and map making
With increasingly large data sets, weak lensing measurements are able to
measure cosmological parameters with ever greater precision. However this
increased accuracy also places greater demands on the statistical tools used to
extract the available information. To date, the majority of lensing analyses
use the two point-statistics of the cosmic shear field. These can either be
studied directly using the two-point correlation function, or in Fourier space,
using the power spectrum. But analyzing weak lensing data inevitably involves
the masking out of regions or example to remove bright stars from the field.
Masking out the stars is common practice but the gaps in the data need proper
handling. In this paper, we show how an inpainting technique allows us to
properly fill in these gaps with only operations, leading to a new
image from which we can compute straight forwardly and with a very good
accuracy both the pow er spectrum and the bispectrum. We propose then a new
method to compute the bispectrum with a polar FFT algorithm, which has the main
advantage of avoiding any interpolation in the Fourier domain. Finally we
propose a new method for dark matter mass map reconstruction from shear
observations which integrates this new inpainting concept. A range of examples
based on 3D N-body simulations illustrates the results.Comment: Final version accepted by MNRAS. The FASTLens software is available
from the following link : http://irfu.cea.fr/Ast/fastlens.software.ph
Structural Variability from Noisy Tomographic Projections
In cryo-electron microscopy, the 3D electric potentials of an ensemble of
molecules are projected along arbitrary viewing directions to yield noisy 2D
images. The volume maps representing these potentials typically exhibit a great
deal of structural variability, which is described by their 3D covariance
matrix. Typically, this covariance matrix is approximately low-rank and can be
used to cluster the volumes or estimate the intrinsic geometry of the
conformation space. We formulate the estimation of this covariance matrix as a
linear inverse problem, yielding a consistent least-squares estimator. For
images of size -by- pixels, we propose an algorithm for calculating this
covariance estimator with computational complexity
, where the condition number
is empirically in the range --. Its efficiency relies on the
observation that the normal equations are equivalent to a deconvolution problem
in 6D. This is then solved by the conjugate gradient method with an appropriate
circulant preconditioner. The result is the first computationally efficient
algorithm for consistent estimation of 3D covariance from noisy projections. It
also compares favorably in runtime with respect to previously proposed
non-consistent estimators. Motivated by the recent success of eigenvalue
shrinkage procedures for high-dimensional covariance matrices, we introduce a
shrinkage procedure that improves accuracy at lower signal-to-noise ratios. We
evaluate our methods on simulated datasets and achieve classification results
comparable to state-of-the-art methods in shorter running time. We also present
results on clustering volumes in an experimental dataset, illustrating the
power of the proposed algorithm for practical determination of structural
variability.Comment: 52 pages, 11 figure
Measurability of kinetic temperature from metal absorption-line spectra formed in chaotic media
We present a new method for recovering the kinetic temperature of the
intervening diffuse gas to an accuracy of 10%. The method is based on the
comparison of unsaturated absorption-line profiles of two species with
different atomic weights. The species are assumed to have the same temperature
and bulk motion within the absorbing region. The computational technique
involves the Fourier transform of the absorption profiles and the consequent
Entropy-Regularized chi^2-Minimization [ERM] to estimate the model parameters.
The procedure is tested using synthetic spectra of CII, SiII and FeII ions. The
comparison with the standard Voigt fitting analysis is performed and it is
shown that the Voigt deconvolution of the complex absorption-line profiles may
result in estimated temperatures which are not physical. We also successfully
analyze Keck telescope spectra of CII1334 and SiII1260 lines observed at the
redshift z = 3.572 toward the quasar Q1937--1009 by Tytler {\it et al.}.Comment: 25 pages, 6 Postscript figures, aaspp4.sty file, submit. Ap
Low Complexity Regularization of Linear Inverse Problems
Inverse problems and regularization theory is a central theme in contemporary
signal processing, where the goal is to reconstruct an unknown signal from
partial indirect, and possibly noisy, measurements of it. A now standard method
for recovering the unknown signal is to solve a convex optimization problem
that enforces some prior knowledge about its structure. This has proved
efficient in many problems routinely encountered in imaging sciences,
statistics and machine learning. This chapter delivers a review of recent
advances in the field where the regularization prior promotes solutions
conforming to some notion of simplicity/low-complexity. These priors encompass
as popular examples sparsity and group sparsity (to capture the compressibility
of natural signals and images), total variation and analysis sparsity (to
promote piecewise regularity), and low-rank (as natural extension of sparsity
to matrix-valued data). Our aim is to provide a unified treatment of all these
regularizations under a single umbrella, namely the theory of partial
smoothness. This framework is very general and accommodates all low-complexity
regularizers just mentioned, as well as many others. Partial smoothness turns
out to be the canonical way to encode low-dimensional models that can be linear
spaces or more general smooth manifolds. This review is intended to serve as a
one stop shop toward the understanding of the theoretical properties of the
so-regularized solutions. It covers a large spectrum including: (i) recovery
guarantees and stability to noise, both in terms of -stability and
model (manifold) identification; (ii) sensitivity analysis to perturbations of
the parameters involved (in particular the observations), with applications to
unbiased risk estimation ; (iii) convergence properties of the forward-backward
proximal splitting scheme, that is particularly well suited to solve the
corresponding large-scale regularized optimization problem
- …