2,287 research outputs found
Transform-based particle filtering for elliptic Bayesian inverse problems
We introduce optimal transport based resampling in adaptive SMC. We consider
elliptic inverse problems of inferring hydraulic conductivity from pressure
measurements. We consider two parametrizations of hydraulic conductivity: by
Gaussian random field, and by a set of scalar (non-)Gaussian distributed
parameters and Gaussian random fields. We show that for scalar parameters
optimal transport based SMC performs comparably to monomial based SMC but for
Gaussian high-dimensional random fields optimal transport based SMC outperforms
monomial based SMC. When comparing to ensemble Kalman inversion with mutation
(EKI), we observe that for Gaussian random fields, optimal transport based SMC
gives comparable or worse performance than EKI depending on the complexity of
the parametrization. For non-Gaussian distributed parameters optimal transport
based SMC outperforms EKI
Multi-scale uncertainty quantification in geostatistical seismic inversion
Geostatistical seismic inversion is commonly used to infer the spatial
distribution of the subsurface petro-elastic properties by perturbing the model
parameter space through iterative stochastic sequential
simulations/co-simulations. The spatial uncertainty of the inferred
petro-elastic properties is represented with the updated a posteriori variance
from an ensemble of the simulated realizations. Within this setting, the
large-scale geological (metaparameters) used to generate the petro-elastic
realizations, such as the spatial correlation model and the global a priori
distribution of the properties of interest, are assumed to be known and
stationary for the entire inversion domain. This assumption leads to
underestimation of the uncertainty associated with the inverted models. We
propose a practical framework to quantify uncertainty of the large-scale
geological parameters in seismic inversion. The framework couples
geostatistical seismic inversion with a stochastic adaptive sampling and
Bayesian inference of the metaparameters to provide a more accurate and
realistic prediction of uncertainty not restricted by heavy assumptions on
large-scale geological parameters. The proposed framework is illustrated with
both synthetic and real case studies. The results show the ability retrieve
more reliable acoustic impedance models with a more adequate uncertainty spread
when compared with conventional geostatistical seismic inversion techniques.
The proposed approach separately account for geological uncertainty at
large-scale (metaparameters) and local scale (trace-by-trace inversion)
Efficient Sequential Monte-Carlo Samplers for Bayesian Inference
In many problems, complex non-Gaussian and/or nonlinear models are required
to accurately describe a physical system of interest. In such cases, Monte
Carlo algorithms are remarkably flexible and extremely powerful approaches to
solve such inference problems. However, in the presence of a high-dimensional
and/or multimodal posterior distribution, it is widely documented that standard
Monte-Carlo techniques could lead to poor performance. In this paper, the study
is focused on a Sequential Monte-Carlo (SMC) sampler framework, a more robust
and efficient Monte Carlo algorithm. Although this approach presents many
advantages over traditional Monte-Carlo methods, the potential of this emergent
technique is however largely underexploited in signal processing. In this work,
we aim at proposing some novel strategies that will improve the efficiency and
facilitate practical implementation of the SMC sampler specifically for signal
processing applications. Firstly, we propose an automatic and adaptive strategy
that selects the sequence of distributions within the SMC sampler that
minimizes the asymptotic variance of the estimator of the posterior
normalization constant. This is critical for performing model selection in
modelling applications in Bayesian signal processing. The second original
contribution we present improves the global efficiency of the SMC sampler by
introducing a novel correction mechanism that allows the use of the particles
generated through all the iterations of the algorithm (instead of only
particles from the last iteration). This is a significant contribution as it
removes the need to discard a large portion of the samples obtained, as is
standard in standard SMC methods. This will improve estimation performance in
practical settings where computational budget is important to consider.Comment: arXiv admin note: text overlap with arXiv:1303.3123 by other author
Adaptive Mixture of Student-t Distributions as a Flexible Candidate Distribution for Efficient Simulation: The R Package AdMit
This paper presents the R package AdMit which provides flexible functions to approximate a certain target distribution and to efficiently generate a sample of random draws from it, given only a kernel of the target density function. The core algorithm consists of the function AdMit which fits an adaptive mixture of Student-t distributions to the density of interest. Then, importance sampling or the independence chain Metropolis-Hastings algorithm is used to obtain quantities of interest for the target density, using the fitted mixture as the importance or candidate density. The estimation procedure is fully automatic and thus avoids the time-consuming and difficult task of tuning a sampling algorithm. The relevance of the package is shown in two examples. The first aims at illustrating in detail the use of the functions provided by the package in a bivariate bimodal distribution. The second shows the relevance of the adaptive mixture procedure through the Bayesian estimation of a mixture of ARCH model fitted to foreign exchange log-returns data. The methodology is compared to standard cases of importance sampling and the Metropolis-Hastings algorithm using a naive candidate and with the Griddy-Gibbs approach.
Computationally Efficient Nonparametric Importance Sampling
The variance reduction established by importance sampling strongly depends on
the choice of the importance sampling distribution. A good choice is often hard
to achieve especially for high-dimensional integration problems. Nonparametric
estimation of the optimal importance sampling distribution (known as
nonparametric importance sampling) is a reasonable alternative to parametric
approaches.In this article nonparametric variants of both the self-normalized
and the unnormalized importance sampling estimator are proposed and
investigated. A common critique on nonparametric importance sampling is the
increased computational burden compared to parametric methods. We solve this
problem to a large degree by utilizing the linear blend frequency polygon
estimator instead of a kernel estimator. Mean square error convergence
properties are investigated leading to recommendations for the efficient
application of nonparametric importance sampling. Particularly, we show that
nonparametric importance sampling asymptotically attains optimal importance
sampling variance. The efficiency of nonparametric importance sampling
algorithms heavily relies on the computational efficiency of the employed
nonparametric estimator. The linear blend frequency polygon outperforms kernel
estimators in terms of certain criteria such as efficient sampling and
evaluation. Furthermore, it is compatible with the inversion method for sample
generation. This allows to combine our algorithms with other variance reduction
techniques such as stratified sampling. Empirical evidence for the usefulness
of the suggested algorithms is obtained by means of three benchmark integration
problems. As an application we estimate the distribution of the queue length of
a spam filter queueing system based on real data.Comment: 29 pages, 7 figure
Bayesian inference for the multivariate skew-normal model: a Population Monte Carlo approach
Frequentist and likelihood methods of inference based on the multivariate
skew-normal model encounter several technical difficulties with this model. In
spite of the popularity of this class of densities, there are no broadly
satisfactory solutions for estimation and testing problems. A general
population Monte Carlo algorithm is proposed which: 1) exploits the latent
structure stochastic representation of skew-normal random variables to provide
a full Bayesian analysis of the model and 2) accounts for the presence of
constraints in the parameter space. The proposed approach can be defined as
weakly informative, since the prior distribution approximates the actual
reference prior for the shape parameter vector. Results are compared with the
existing classical solutions and the practical implementation of the algorithm
is illustrated via a simulation study and a real data example. A generalization
to the matrix variate regression model with skew-normal error is also
presented
- …