8,595 research outputs found
Calculating partial expected value of perfect information via Monte Carlo sampling algorithms
Partial expected value of perfect information (EVPI) calculations can quantify the value of learning about particular subsets of uncertain parameters in decision models. Published case studies have used different computational approaches. This article examines the computation of partial EVPI estimates via Monte Carlo sampling algorithms. The mathematical definition shows 2 nested expectations, which must be evaluated separately because of the need to compute a maximum between them. A generalized Monte Carlo sampling algorithm uses nested simulation with an outer loop to sample parameters of interest and, conditional upon these, an inner loop to sample remaining uncertain parameters. Alternative computation methods and shortcut algorithms are discussed and mathematical conditions for their use considered. Maxima of Monte Carlo estimates of expectations are biased upward, and the authors show that the use of small samples results in biased EVPI estimates. Three case studies illustrate 1) the bias due to maximization and also the inaccuracy of shortcut algorithms 2) when correlated variables are present and 3) when there is nonlinearity in net benefit functions. If relatively small correlation or nonlinearity is present, then the shortcut algorithm can be substantially inaccurate. Empirical investigation of the numbers of Monte Carlo samples suggests that fewer samples on the outer level and more on the inner level could be efficient and that relatively small numbers of samples can sometimes be used. Several remaining areas for methodological development are set out. A wider application of partial EVPI is recommended both for greater understanding of decision uncertainty and for analyzing research priorities
Bayesian comparison of latent variable models: Conditional vs marginal likelihoods
Typical Bayesian methods for models with latent variables (or random effects)
involve directly sampling the latent variables along with the model parameters.
In high-level software code for model definitions (using, e.g., BUGS, JAGS,
Stan), the likelihood is therefore specified as conditional on the latent
variables. This can lead researchers to perform model comparisons via
conditional likelihoods, where the latent variables are considered model
parameters. In other settings, however, typical model comparisons involve
marginal likelihoods where the latent variables are integrated out. This
distinction is often overlooked despite the fact that it can have a large
impact on the comparisons of interest. In this paper, we clarify and illustrate
these issues, focusing on the comparison of conditional and marginal Deviance
Information Criteria (DICs) and Watanabe-Akaike Information Criteria (WAICs) in
psychometric modeling. The conditional/marginal distinction corresponds to
whether the model should be predictive for the clusters that are in the data or
for new clusters (where "clusters" typically correspond to higher-level units
like people or schools). Correspondingly, we show that marginal WAIC
corresponds to leave-one-cluster out (LOcO) cross-validation, whereas
conditional WAIC corresponds to leave-one-unit out (LOuO). These results lead
to recommendations on the general application of the criteria to models with
latent variables.Comment: Manuscript in press at Psychometrika; 31 pages, 8 figure
Split Sampling: Expectations, Normalisation and Rare Events
In this paper we develop a methodology that we call split sampling methods to
estimate high dimensional expectations and rare event probabilities. Split
sampling uses an auxiliary variable MCMC simulation and expresses the
expectation of interest as an integrated set of rare event probabilities. We
derive our estimator from a Rao-Blackwellised estimate of a marginal auxiliary
variable distribution. We illustrate our method with two applications. First,
we compute a shortest network path rare event probability and compare our
method to estimation to a cross entropy approach. Then, we compute a
normalisation constant of a high dimensional mixture of Gaussians and compare
our estimate to one based on nested sampling. We discuss the relationship
between our method and other alternatives such as the product of conditional
probability estimator and importance sampling. The methods developed here are
available in the R package: SplitSampling
Decision-making under uncertainty: using MLMC for efficient estimation of EVPPI
In this paper we develop a very efficient approach to the Monte Carlo
estimation of the expected value of partial perfect information (EVPPI) that
measures the average benefit of knowing the value of a subset of uncertain
parameters involved in a decision model. The calculation of EVPPI is inherently
a nested expectation problem, with an outer expectation with respect to one
random variable and an inner conditional expectation with respect to the
other random variable . We tackle this problem by using a Multilevel Monte
Carlo (MLMC) method (Giles 2008) in which the number of inner samples for
increases geometrically with level, so that the accuracy of estimating the
inner conditional expectation improves and the cost also increases with level.
We construct an antithetic MLMC estimator and provide sufficient assumptions on
a decision model under which the antithetic property of the estimator is well
exploited, and consequently a root-mean-square accuracy of can be
achieved at a cost of . Numerical results confirm the
considerable computational savings compared to the standard, nested Monte Carlo
method for some simple testcases and a more realistic medical application
A machine learning approach to portfolio pricing and risk management for high-dimensional problems
We present a general framework for portfolio risk management in discrete
time, based on a replicating martingale. This martingale is learned from a
finite sample in a supervised setting. The model learns the features necessary
for an effective low-dimensional representation, overcoming the curse of
dimensionality common to function approximation in high-dimensional spaces. We
show results based on polynomial and neural network bases. Both offer superior
results to naive Monte Carlo methods and other existing methods like
least-squares Monte Carlo and replicating portfolios.Comment: 30 pages (main), 10 pages (appendix), 3 figures, 22 table
On Nesting Monte Carlo Estimators
Many problems in machine learning and statistics involve nested expectations
and thus do not permit conventional Monte Carlo (MC) estimation. For such
problems, one must nest estimators, such that terms in an outer estimator
themselves involve calculation of a separate, nested, estimation. We
investigate the statistical implications of nesting MC estimators, including
cases of multiple levels of nesting, and establish the conditions under which
they converge. We derive corresponding rates of convergence and provide
empirical evidence that these rates are observed in practice. We further
establish a number of pitfalls that can arise from naive nesting of MC
estimators, provide guidelines about how these can be avoided, and lay out
novel methods for reformulating certain classes of nested expectation problems
into single expectations, leading to improved convergence rates. We demonstrate
the applicability of our work by using our results to develop a new estimator
for discrete Bayesian experimental design problems and derive error bounds for
a class of variational objectives.Comment: To appear at International Conference on Machine Learning 201
- …