8 research outputs found
HINT: Hierarchical Invertible Neural Transport for Density Estimation and Bayesian Inference
A large proportion of recent invertible neural architectures is based on a
coupling block design. It operates by dividing incoming variables into two
sub-spaces, one of which parameterizes an easily invertible (usually affine)
transformation that is applied to the other. While the Jacobian of such a
transformation is triangular, it is very sparse and thus may lack
expressiveness. This work presents a simple remedy by noting that (affine)
coupling can be repeated recursively within the resulting sub-spaces, leading
to an efficiently invertible block with dense triangular Jacobian. By
formulating our recursive coupling scheme via a hierarchical architecture, HINT
allows sampling from a joint distribution p(y,x) and the corresponding
posterior p(x|y) using a single invertible network. We demonstrate the power of
our method for density estimation and Bayesian inference on a novel data set of
2D shapes in Fourier parameterization, which enables consistent visualization
of samples for different dimensionalities
Continuous Level Monte Carlo and Sample-Adaptive Model Hierarchies
In this paper, we present a generalization of the multilevel Monte Carlo (MLMC) method to a setting where the level parameter is a continuous variable. This continuous level Monte Carlo (CLMC) estimator provides a natural framework in PDE applications to adapt the model hierarchy to each sample. In addition, it can be made unbiased with respect to the expected value of the true quantity of interest provided the quantity of interest converges sufficiently fast. The practical implementation of the CLMC estimator is based on interpolating actual evaluations of the quantity of interest at a finite number of resolutions. As our new level parameter, we use the logarithm of a goal-oriented finite element error estimator for the accuracy of the quantity of interest. We prove the unbiasedness, as well as a complexity theorem that shows the same rate of complexity for CLMC as for MLMC. Finally, we provide some numerical evidence to support our theoretical results, by successfully testing CLMC on a standard PDE test problem. The numerical experiments demonstrate clear gains for samplewise adaptive refinement strategies over uniform refinements.</p
A Stein variational Newton method
Stein variational gradient descent (SVGD) was recently proposed as a general
purpose nonparametric variational inference algorithm [Liu & Wang, NIPS 2016]:
it minimizes the Kullback-Leibler divergence between the target distribution
and its approximation by implementing a form of functional gradient descent on
a reproducing kernel Hilbert space. In this paper, we accelerate and generalize
the SVGD algorithm by including second-order information, thereby approximating
a Newton-like iteration in function space. We also show how second-order
information can lead to more effective choices of kernel. We observe
significant computational gains over the original SVGD algorithm in multiple
test cases.Comment: 18 pages, 7 figure
Multilevel Dimension-Independent Likelihood-Informed MCMC for Large-Scale Inverse Problems
We present a non-trivial integration of dimension-independent
likelihood-informed (DILI) MCMC (Cui, Law, Marzouk, 2016) and the multilevel
MCMC (Dodwell et al., 2015) to explore the hierarchy of posterior
distributions. This integration offers several advantages: First, DILI-MCMC
employs an intrinsic likelihood-informed subspace (LIS) (Cui et al., 2014) --
which involves a number of forward and adjoint model simulations -- to design
accelerated operator-weighted proposals. By exploiting the multilevel structure
of the discretised parameters and discretised forward models, we design a
Rayleigh-Ritz procedure to significantly reduce the computational effort in
building the LIS and operating with DILI proposals. Second, the resulting
DILI-MCMC can drastically improve the sampling efficiency of MCMC at each
level, and hence reduce the integration error of the multilevel algorithm for
fixed CPU time. To be able to fully exploit the power of multilevel MCMC and to
reduce the dependencies of samples on different levels for a parallel
implementation, we also suggest a new pooling strategy for allocating
computational resources across different levels and constructing Markov chains
at higher levels conditioned on those simulated on lower levels. Numerical
results confirm the improved computational efficiency of the multilevel DILI
approach
Continuous Level Monte Carlo and Sample-Adaptive Model Hierarchies
This is the final version. Available from SIAM via the DOI in this record.In this paper, we present a generalisation of the Multilevel Monte Carlo (MLMC) method to a setting where the level parameter is a continuous variable. This Continuous Level Monte Carlo (CLMC) estimator provides a natural framework in PDE applications to adapt the model hierarchy to each sample. In addition, it can be made unbiased with respect to the expected value of the true quantity of interest provided the quantity of interest converges sufficiently fast. The practical implementation of the CLMC estimator is based on interpolating actual evaluations of the quantity of interest at a finite number of resolutions. As our new level parameter, we use the logarithm of a goal-oriented finite element error estimator for the accuracy of the quantity of interest. We prove the unbiasedness, as well as a complexity theorem that shows the same rate of complexity for CLMC as for MLMC. Finally, we provide some numerical evidence to support our theoretical results, by successfully testing CLMC on a standard PDE test problem. The numerical experiments demonstrate clear gains for sample-wise adaptive refinement strategies over uniform refinements