5,900 research outputs found
Sequential Monte Carlo EM for multivariate probit models
Multivariate probit models (MPM) have the appealing feature of capturing some
of the dependence structure between the components of multidimensional binary
responses. The key for the dependence modelling is the covariance matrix of an
underlying latent multivariate Gaussian. Most approaches to MLE in multivariate
probit regression rely on MCEM algorithms to avoid computationally intensive
evaluations of multivariate normal orthant probabilities. As an alternative to
the much used Gibbs sampler a new SMC sampler for truncated multivariate
normals is proposed. The algorithm proceeds in two stages where samples are
first drawn from truncated multivariate Student distributions and then
further evolved towards a Gaussian. The sampler is then embedded in a MCEM
algorithm. The sequential nature of SMC methods can be exploited to design a
fully sequential version of the EM, where the samples are simply updated from
one iteration to the next rather than resampled from scratch. Recycling the
samples in this manner significantly reduces the computational cost. An
alternative view of the standard conditional maximisation step provides the
basis for an iterative procedure to fully perform the maximisation needed in
the EM algorithm. The identifiability of MPM is also thoroughly discussed. In
particular, the likelihood invariance can be embedded in the EM algorithm to
ensure that constrained and unconstrained maximisation are equivalent. A simple
iterative procedure is then derived for either maximisation which takes
effectively no computational time. The method is validated by applying it to
the widely analysed Six Cities dataset and on a higher dimensional simulated
example. Previous approaches to the Six Cities overly restrict the parameter
space but, by considering the correct invariance, the maximum likelihood is
quite naturally improved when treating the full unrestricted model.Comment: 26 pages, 2 figures. In press, Computational Statistics & Data
Analysi
Computation of Gaussian orthant probabilities in high dimension
We study the computation of Gaussian orthant probabilities, i.e. the
probability that a Gaussian falls inside a quadrant. The
Geweke-Hajivassiliou-Keane (GHK) algorithm [Genz, 1992; Geweke, 1991;
Hajivassiliou et al., 1996; Keane, 1993], is currently used for integrals of
dimension greater than 10. In this paper we show that for Markovian covariances
GHK can be interpreted as the estimator of the normalizing constant of a state
space model using sequential importance sampling (SIS). We show for an AR(1)
the variance of the GHK, properly normalized, diverges exponentially fast with
the dimension. As an improvement we propose using a particle filter (PF). We
then generalize this idea to arbitrary covariance matrices using Sequential
Monte Carlo (SMC) with properly tailored MCMC moves. We show empirically that
this can lead to drastic improvements on currently used algorithms. We also
extend the framework to orthants of mixture of Gaussians (Student, Cauchy
etc.), and to the simulation of truncated Gaussians
Open TURNS: An industrial software for uncertainty quantification in simulation
The needs to assess robust performances for complex systems and to answer
tighter regulatory processes (security, safety, environmental control, and
health impacts, etc.) have led to the emergence of a new industrial simulation
challenge: to take uncertainties into account when dealing with complex
numerical simulation frameworks. Therefore, a generic methodology has emerged
from the joint effort of several industrial companies and academic
institutions. EDF R&D, Airbus Group and Phimeca Engineering started a
collaboration at the beginning of 2005, joined by IMACS in 2014, for the
development of an Open Source software platform dedicated to uncertainty
propagation by probabilistic methods, named OpenTURNS for Open source Treatment
of Uncertainty, Risk 'N Statistics. OpenTURNS addresses the specific industrial
challenges attached to uncertainties, which are transparency, genericity,
modularity and multi-accessibility. This paper focuses on OpenTURNS and
presents its main features: openTURNS is an open source software under the LGPL
license, that presents itself as a C++ library and a Python TUI, and which
works under Linux and Windows environment. All the methodological tools are
described in the different sections of this paper: uncertainty quantification,
uncertainty propagation, sensitivity analysis and metamodeling. A section also
explains the generic wrappers way to link openTURNS to any external code. The
paper illustrates as much as possible the methodological tools on an
educational example that simulates the height of a river and compares it to the
height of a dyke that protects industrial facilities. At last, it gives an
overview of the main developments planned for the next few years
Stochastic volatility models for ordinal valued time series with application to finance
In this paper we introduce two stochastic volatility models where the response variable takes on only finite many ordered values. Corresponding time series occur in high-frequency finance when the stocks are traded on a coarse grid. For parameter estimation we develop an e±cient Grouped Move Multigrid Monte Carlo (GM-MGMC) sampler. We apply both models to price changes of the IBM stock in January, 2001 at the NYSE. Dependencies of the price change process on covariates are quantified and compared with theoretical considerations on such processes. We also investigate whether this data set requires modeling with a heavy-tailed Student-t distribution
- …