8 research outputs found
Gaussian process hyper-parameter estimation using parallel asymptotically independent Markov sampling
Gaussian process emulators of computationally expensive computer codes
provide fast statistical approximations to model physical processes. The
training of these surrogates depends on the set of design points chosen to run
the simulator. Due to computational cost, such training set is bound to be
limited and quantifying the resulting uncertainty in the hyper-parameters of
the emulator by uni-modal distributions is likely to induce bias. In order to
quantify this uncertainty, this paper proposes a computationally efficient
sampler based on an extension of Asymptotically Independent Markov Sampling, a
recently developed algorithm for Bayesian inference. Structural uncertainty of
the emulator is obtained as a by-product of the Bayesian treatment of the
hyper-parameters. Additionally, the user can choose to perform stochastic
optimisation to sample from a neighbourhood of the Maximum a Posteriori
estimate, even in the presence of multimodality. Model uncertainty is also
acknowledged through numerical stabilisation measures by including a nugget
term in the formulation of the probability model. The efficiency of the
proposed sampler is illustrated in examples where multi-modal distributions are
encountered. For the purpose of reproducibility, further development, and use
in other applications the code used to generate the examples is freely
available for download at https://github.com/agarbuno/paims_codesComment: Computational Statistics \& Data Analysis, Volume 103, November 201
Affine invariant interacting Langevin dynamics for Bayesian inference
We propose a computational method (with acronym ALDI) for sampling from a given target distribution based on first-order (overdamped) Langevin dynamics which satisfies the property of affine invariance. The central idea of ALDI is to run an ensemble of particles with their empirical covariance serving as a preconditioner for their underlying Langevin dynamics. ALDI does not require taking the inverse or square root of the empirical covariance matrix, which enables application to high-dimensional sampling problems. The theoretical properties of ALDI are studied in terms of non-degeneracy and ergodicity. Furthermore, we study its connections to diffusions on Riemannian manifolds and Wasserstein gradient flows.
Bayesian inference serves as a main application area for ALDI. In case of a forward problem with additive Gaussian measurement errors, ALDI allows for a gradient-free implementation in the spirit of the ensemble Kalman filter. A computational comparison between gradient-free and gradient-based ALDI is provided for a PDE constrained Bayesian inverse problem
Stochastic Methods for Emulation, Calibration and Reliability Analysis of Engineering Models
This dissertation examines the use of non-parametric Bayesian methods and advanced Monte Carlo algorithms for the emulation and reliability analysis of complex engineering computations. Firstly, the problem lies in the reduction of the computational cost of such models and the generation of posterior samples for the Gaussian Process’ (GP) hyperparameters. In a GP, as the flexibility of the mechanism to induce correlations among training points increases, the number of hyperparameters increases as well. This leads to multimodal posterior distributions. Typical variants of MCMC samplers are not designed to overcome multimodality. Maximum posterior estimates of hyperparameters, on the other hand, do not guarantee a global optimiser. This presents a challenge when emulating expensive simulators in light of small data. Thus, new MCMC algorithms are presented which allow the use of full Bayesian emulators by sampling from their respective multimodal posteriors. Secondly, in order for these complex models to be reliable, they need to be robustly calibrated to experimental data. History matching solves the calibration problem by discarding regions of input parameters space. This allows one to determine which configurations are likely to replicate the observed data. In particular, the GP surrogate model’s probabilistic statements are exploited, and the data assimilation process is improved. Thirdly, as sampling- based methods are increasingly being used in engineering, variants of sampling algorithms in other engineering tasks are studied, that is reliability-based methods. Several new algorithms to solve these three fundamental problems are proposed, developed and tested in both illustrative examples and industrial-scale models
Transitional annealed adaptive slice sampling for Gaussian process hyper-parameter estimation
Surrogate models have become ubiquitous in science and engineering for their capability of emulating expensive computer codes, necessary to model and investigate complex phenomena. Bayesian emulators based on Gaussian processes adequately quantify the uncertainty that results from the cost of the original simulator, and thus the inability to evaluate it on the whole input space. However, it is common in the literature that only a partial Bayesian analysis is carried out, whereby the underlying hyper-parameters are estimated via gradient-free optimization or genetic algorithms, to name a few methods. On the other hand, maximum a posteriori (MAP) estimation could discard important regions of the hyper-parameter space. In this paper, we carry out a more complete Bayesian inference, that combines Slice Sampling with some recently developed sequential Monte Carlo samplers. The resulting algorithm improves the mixing in the sampling through the delayed-rejection nature of Slice Sampling, the inclusion of an annealing scheme akin to Asymptotically Independent Markov Sampling and parallelization via transitional Markov chain Monte Carlo. Examples related to the estimation of Gaussian process hyper-parameters are presented. For the purpose of reproducibility, further development, and use in other applications, the code to generate the examples in this paper is freely available for download at http://github.com/agarbuno/ta2s2_codes
Calibration and Uncertainty Quantification of Convective Parameters in an Idealized GCM
Parameters in climate models are usually calibrated manually, exploiting only
small subsets of the available data. This precludes both optimal calibration
and quantification of uncertainties. Traditional Bayesian calibration methods
that allow uncertainty quantification are too expensive for climate models;
they are also not robust in the presence of internal climate variability. For
example, Markov chain Monte Carlo (MCMC) methods typically require
model runs and are sensitive to internal variability noise, rendering them
infeasible for climate models. Here we demonstrate an approach to model
calibration and uncertainty quantification that requires only model
runs and can accommodate internal climate variability. The approach consists of
three stages: (i) a calibration stage uses variants of ensemble Kalman
inversion to calibrate a model by minimizing mismatches between model and data
statistics; (ii) an emulation stage emulates the parameter-to-data map with
Gaussian processes (GP), using the model runs in the calibration stage for
training; (iii) a sampling stage approximates the Bayesian posterior
distributions by sampling the GP emulator with MCMC. We demonstrate the
feasibility and computational efficiency of this calibrate-emulate-sample (CES)
approach in a perfect-model setting. Using an idealized general circulation
model, we estimate parameters in a simple convection scheme from synthetic data
generated with the model. The CES approach generates probability distributions
of the parameters that are good approximations of the Bayesian posteriors, at a
fraction of the computational cost usually required to obtain them. Sampling
from this approximate posterior allows the generation of climate predictions
with quantified parametric uncertainties
Transitional annealed adaptive slice sampling for Gaussian process hyper-parameter estimation
Surrogate models have become ubiquitous in science and engineering for their capability of emulating expensive computer codes, necessary to model and investigate complex phenomena. Bayesian emulators based on Gaussian processes adequately quantify the uncertainty that results from the cost of the original simulator, and thus the inability to evaluate it on the whole input space. However, it is common in the literature that only a partial Bayesian analysis is carried out, whereby the underlying hyper-parameters are estimated via gradient-free optimization or genetic algorithms, to name a few methods. On the other hand, maximum a posteriori (MAP) estimation could discard important regions of the hyper-parameter space. In this paper, we carry out a more complete Bayesian inference, that combines Slice Sampling with some recently developed sequential Monte Carlo samplers. The resulting algorithm improves the mixing in the sampling through the delayed-rejection nature of Slice Sampling, the inclusion of an annealing scheme akin to Asymptotically Independent Markov Sampling and parallelization via transitional Markov chain Monte Carlo. Examples related to the estimation of Gaussian process hyper-parameters are presented. For the purpose of reproducibility, further development, and use in other applications, the code to generate the examples in this paper is freely available for download at http://github.com/agarbuno/ta2s2_codes
Bayesian updating and model class selection with Subset Simulation
Identifying the parameters of a model and rating competitive models based on
measured data has been among the most important but challenging topics in
modern science and engineering, with great potential of application in
structural system identification, updating and development of high fidelity
models. These problems in principle can be tackled using a Bayesian
probabilistic approach, where the parameters to be identified are treated as
uncertain and their inference information are given in terms of their posterior
(i.e., given data) probability distribution. For complex models encountered in
applications, efficient computational tools robust to the number of uncertain
parameters in the problem are required for computing the posterior statistics,
which can generally be formulated as a multi-dimensional integral over the
space of the uncertain parameters. Subset Simulation (SuS) has been developed
for solving reliability problems involving complex systems and it is found to
be robust to the number of uncertain parameters. An analogy has been recently
established between a Bayesian updating problem and a reliability problem,
which opens up the possibility of efficient solution by SuS. The formulation,
called BUS (Bayesian Updating with Structural reliability methods), is based on
conventional rejection principle. Its theoretical correctness and efficiency
requires the prudent choice of a multiplier, which has remained an open
question. Motivated by the choice of the multiplier and its philosophical role,
this paper presents a study of BUS. The work leads to a revised formulation
that resolves the issues regarding the multiplier so that SuS can be
implemented without knowing the multiplier. Examples are presented to
illustrate the theory and applications