1,132 research outputs found
Particle Metropolis-Hastings using gradient and Hessian information
Particle Metropolis-Hastings (PMH) allows for Bayesian parameter inference in
nonlinear state space models by combining Markov chain Monte Carlo (MCMC) and
particle filtering. The latter is used to estimate the intractable likelihood.
In its original formulation, PMH makes use of a marginal MCMC proposal for the
parameters, typically a Gaussian random walk. However, this can lead to a poor
exploration of the parameter space and an inefficient use of the generated
particles.
We propose a number of alternative versions of PMH that incorporate gradient
and Hessian information about the posterior into the proposal. This information
is more or less obtained as a byproduct of the likelihood estimation. Indeed,
we show how to estimate the required information using a fixed-lag particle
smoother, with a computational cost growing linearly in the number of
particles. We conclude that the proposed methods can: (i) decrease the length
of the burn-in phase, (ii) increase the mixing of the Markov chain at the
stationary phase, and (iii) make the proposal distribution scale invariant
which simplifies tuning.Comment: 27 pages, 5 figures, 2 tables. The final publication is available at
Springer via: http://dx.doi.org/10.1007/s11222-014-9510-
Estimating the granularity coefficient of a Potts-Markov random field within an MCMC algorithm
This paper addresses the problem of estimating the Potts parameter B jointly
with the unknown parameters of a Bayesian model within a Markov chain Monte
Carlo (MCMC) algorithm. Standard MCMC methods cannot be applied to this problem
because performing inference on B requires computing the intractable
normalizing constant of the Potts model. In the proposed MCMC method the
estimation of B is conducted using a likelihood-free Metropolis-Hastings
algorithm. Experimental results obtained for synthetic data show that
estimating B jointly with the other unknown parameters leads to estimation
results that are as good as those obtained with the actual value of B. On the
other hand, assuming that the value of B is known can degrade estimation
performance significantly if this value is incorrect. To illustrate the
interest of this method, the proposed algorithm is successfully applied to real
bidimensional SAR and tridimensional ultrasound images
Accelerating MCMC Algorithms
Markov chain Monte Carlo algorithms are used to simulate from complex
statistical distributions by way of a local exploration of these distributions.
This local feature avoids heavy requests on understanding the nature of the
target, but it also potentially induces a lengthy exploration of this target,
with a requirement on the number of simulations that grows with the dimension
of the problem and with the complexity of the data behind it. Several
techniques are available towards accelerating the convergence of these Monte
Carlo algorithms, either at the exploration level (as in tempering, Hamiltonian
Monte Carlo and partly deterministic methods) or at the exploitation level
(with Rao-Blackwellisation and scalable methods).Comment: This is a survey paper, submitted WIREs Computational Statistics, to
with 6 figure
Importance Tempering
Simulated tempering (ST) is an established Markov chain Monte Carlo (MCMC)
method for sampling from a multimodal density . Typically, ST
involves introducing an auxiliary variable taking values in a finite subset
of and indexing a set of tempered distributions, say . In this case, small values of encourage better
mixing, but samples from are only obtained when the joint chain for
reaches . However, the entire chain can be used to estimate
expectations under of functions of interest, provided that importance
sampling (IS) weights are calculated. Unfortunately this method, which we call
importance tempering (IT), can disappoint. This is partly because the most
immediately obvious implementation is na\"ive and can lead to high variance
estimators. We derive a new optimal method for combining multiple IS estimators
and prove that the resulting estimator has a highly desirable property related
to the notion of effective sample size. We briefly report on the success of the
optimal combination in two modelling scenarios requiring reversible-jump MCMC,
where the na\"ive approach fails.Comment: 16 pages, 2 tables, significantly shortened from version 4 in
response to referee comments, to appear in Statistics and Computin
Patterns of Scalable Bayesian Inference
Datasets are growing not just in size but in complexity, creating a demand
for rich models and quantification of uncertainty. Bayesian methods are an
excellent fit for this demand, but scaling Bayesian inference is a challenge.
In response to this challenge, there has been considerable recent work based on
varying assumptions about model structure, underlying computational resources,
and the importance of asymptotic correctness. As a result, there is a zoo of
ideas with few clear overarching principles.
In this paper, we seek to identify unifying principles, patterns, and
intuitions for scaling Bayesian inference. We review existing work on utilizing
modern computing resources with both MCMC and variational approximation
techniques. From this taxonomy of ideas, we characterize the general principles
that have proven successful for designing scalable inference procedures and
comment on the path forward
- …