262 research outputs found

    Monte Carlo likelihood inference for missing data models

    Full text link
    We describe a Monte Carlo method to approximate the maximum likelihood estimate (MLE), when there are missing data and the observed data likelihood is not available in closed form. This method uses simulated missing data that are independent and identically distributed and independent of the observed data. Our Monte Carlo approximation to the MLE is a consistent and asymptotically normal estimate of the minimizer θ∗\theta^* of the Kullback--Leibler information, as both Monte Carlo and observed data sample sizes go to infinity simultaneously. Plug-in estimates of the asymptotic variance are provided for constructing confidence regions for θ∗\theta^*. We give Logit--Normal generalized linear mixed model examples, calculated using an R package.Comment: Published at http://dx.doi.org/10.1214/009053606000001389 in the Annals of Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical Statistics (http://www.imstat.org

    Surreal Time and Ultratasks

    Get PDF
    This paper suggests that time could have a much richer mathematical structure than that of the real numbers. Clark & Read (1984) argue that a hypertask (uncountably many tasks done in a finite length of time) cannot be performed. Assuming that time takes values in the real numbers, we give a trivial proof of this. If we instead take the surreal numbers as a model of time, then not only are hypertasks possible but so is an ultratask (a sequence which includes one task done for each ordinal number—thus a proper class of them). We argue that the surreal numbers are in some respects a better model of the temporal continuum than the real numbers as defined in mainstream mathematics, and that surreal time and hypertasks are mathematically possible

    Ridge Fusion in Statistical Learning

    Full text link
    We propose a penalized likelihood method to jointly estimate multiple precision matrices for use in quadratic discriminant analysis and model based clustering. A ridge penalty and a ridge fusion penalty are used to introduce shrinkage and promote similarity between precision matrix estimates. Block-wise coordinate descent is used for optimization, and validation likelihood is used for tuning parameter selection. Our method is applied in quadratic discriminant analysis and semi-supervised model based clustering.Comment: 24 pages and 9 tables, 3 figure

    Likelihood Ration Tests and Inequality Contraints

    Get PDF
    1 online resource (PDF, 28 pages

    Computation for the Introduction to MCMC Chapter Handbook of Markov Chain Monte Carlo

    Get PDF
    1 online resource (PDF, 11 pages
    • …
    corecore