12 research outputs found
Random forward models and log-likelihoods in Bayesian inverse problems
We consider the use of randomised forward models and log-likelihoods within
the Bayesian approach to inverse problems. Such random approximations to the
exact forward model or log-likelihood arise naturally when a computationally
expensive model is approximated using a cheaper stochastic surrogate, as in
Gaussian process emulation (kriging), or in the field of probabilistic
numerical methods. We show that the Hellinger distance between the exact and
approximate Bayesian posteriors is bounded by moments of the difference between
the true and approximate log-likelihoods. Example applications of these
stability results are given for randomised misfit models in large data
applications and the probabilistic solution of ordinary differential equations.Comment: 25 page
Strong convergence rates of probabilistic integrators for ordinary differential equations
Probabilistic integration of a continuous dynamical system is a way of
systematically introducing model error, at scales no larger than errors
introduced by standard numerical discretisation, in order to enable thorough
exploration of possible responses of the system to inputs. It is thus a
potentially useful approach in a number of applications such as forward
uncertainty quantification, inverse problems, and data assimilation. We extend
the convergence analysis of probabilistic integrators for deterministic
ordinary differential equations, as proposed by Conrad et al.\ (\textit{Stat.\
Comput.}, 2017), to establish mean-square convergence in the uniform norm on
discrete- or continuous-time solutions under relaxed regularity assumptions on
the driving vector fields and their induced flows. Specifically, we show that
randomised high-order integrators for globally Lipschitz flows and randomised
Euler integrators for dissipative vector fields with polynomially-bounded local
Lipschitz constants all have the same mean-square convergence rate as their
deterministic counterparts, provided that the variance of the integration noise
is not of higher order than the corresponding deterministic integrator. These
and similar results are proven for probabilistic integrators where the random
perturbations may be state-dependent, non-Gaussian, or non-centred random
variables.Comment: 25 page
Convergence of Gaussian Process Regression with Estimated Hyper-parameters and Applications in Bayesian Inverse Problems
This work is concerned with the convergence of Gaussian process regression. A
particular focus is on hierarchical Gaussian process regression, where
hyper-parameters appearing in the mean and covariance structure of the Gaussian
process emulator are a-priori unknown, and are learnt from the data, along with
the posterior mean and covariance. We work in the framework of empirical Bayes,
where a point estimate of the hyper-parameters is computed, using the data, and
then used within the standard Gaussian process prior to posterior update. We
provide a convergence analysis that (i) holds for any continuous function
to be emulated; and (ii) shows that convergence of Gaussian process regression
is unaffected by the additional learning of hyper-parameters from data, and is
guaranteed in a wide range of scenarios. As the primary motivation for the work
is the use of Gaussian process regression to approximate the data likelihood in
Bayesian inverse problems, we provide a bound on the error introduced in the
Bayesian posterior distribution in this context
A probabilistic finite element method based on random meshes: Error estimators and Bayesian inverse problems
We present a novel probabilistic finite element method (FEM) for the solution
and uncertainty quantification of elliptic partial differential equations based
on random meshes, which we call random mesh FEM (RM-FEM). Our methodology
allows to introduce a probability measure on standard piecewise linear FEM. We
present a posteriori error estimators based uniquely on probabilistic
information. A series of numerical experiments illustrates the potential of the
RM-FEM for error estimation and validates our analysis. We furthermore
demonstrate how employing the RM-FEM enhances the quality of the solution of
Bayesian inverse problems, thus allowing a better quantification of numerical
errors in pipelines of computations
Randomised one-step time integration methods for deterministic operator differential equations
Uncertainty quantification plays an important role in problems that involve inferring a parameter of an initial value problem from observations of the solution. Conrad et al. (Stat Comput 27(4):1065–1082, 2017) proposed randomisation of deterministic time integration methods as a strategy for quantifying uncertainty due to the unknown time discretisation error. We consider this strategy for systems that are described by deterministic, possibly time-dependent operator differential equations defined on a Banach space or a Gelfand triple. Our main results are strong error bounds on the random trajectories measured in Orlicz norms, proven under a weaker assumption on the local truncation error of the underlying deterministic time integration method. Our analysis establishes the theoretical validity of randomised time integration for differential equations in infinite-dimensional settings