8,302 research outputs found

    Iterative Updating of Model Error for Bayesian Inversion

    Get PDF
    In computational inverse problems, it is common that a detailed and accurate forward model is approximated by a computationally less challenging substitute. The model reduction may be necessary to meet constraints in computing time when optimization algorithms are used to find a single estimate, or to speed up Markov chain Monte Carlo (MCMC) calculations in the Bayesian framework. The use of an approximate model introduces a discrepancy, or modeling error, that may have a detrimental effect on the solution of the ill-posed inverse problem, or it may severely distort the estimate of the posterior distribution. In the Bayesian paradigm, the modeling error can be considered as a random variable, and by using an estimate of the probability distribution of the unknown, one may estimate the probability distribution of the modeling error and incorporate it into the inversion. We introduce an algorithm which iterates this idea to update the distribution of the model error, leading to a sequence of posterior distributions that are demonstrated empirically to capture the underlying truth with increasing accuracy. Since the algorithm is not based on rejections, it requires only limited full model evaluations. We show analytically that, in the linear Gaussian case, the algorithm converges geometrically fast with respect to the number of iterations. For more general models, we introduce particle approximations of the iteratively generated sequence of distributions; we also prove that each element of the sequence converges in the large particle limit. We show numerically that, as in the linear case, rapid convergence occurs with respect to the number of iterations. Additionally, we show through computed examples that point estimates obtained from this iterative algorithm are superior to those obtained by neglecting the model error.Comment: 39 pages, 9 figure

    Bayesian peak-bagging of solar-like oscillators using MCMC: A comprehensive guide

    Full text link
    Context: Asteroseismology has entered a new era with the advent of the NASA Kepler mission. Long and continuous photometric observations of unprecedented quality are now available which have stimulated the development of a number of suites of innovative analysis tools. Aims: The power spectra of solar-like oscillations are an inexhaustible source of information on stellar structure and evolution. Robust methods are hence needed in order to infer both individual oscillation mode parameters and parameters describing non-resonant features, thus making a seismic interpretation possible. Methods: We present a comprehensive guide to the implementation of a Bayesian peak-bagging tool that employs a Markov chain Monte Carlo (MCMC). Besides making it possible to incorporate relevant prior information through Bayes' theorem, this tool also allows one to obtain the marginal probability density function for each of the fitted parameters. We apply this tool to a couple of recent asteroseismic data sets, namely, to CoRoT observations of HD 49933 and to ground-based observations made during a campaign devoted to Procyon. Results: The developed method performs remarkably well at constraining not only in the traditional case of extracting oscillation frequencies, but also when pushing the limit where traditional methods have difficulties. Moreover it provides an rigorous way of comparing competing models, such as the ridge identifications, against the asteroseismic data.Comment: Accepted for publication in A&

    Cosmology of neutrinos and extra light particles after WMAP3

    Full text link
    We study how present data probe standard and non-standard properties of neutrinos and the possible existence of new light particles, freely-streaming or interacting, among themselves or with neutrinos. Our results include: sum m_nu < 0.40 eV at 99.9% C.L.; that extra massless particles have abundance Delta N_nu = 2 pm 1 if freely-streaming and Delta N_nu = 0 pm 1.3 if interacting; that 3 interacting neutrinos are disfavored at about 4 sigma. We investigate the robustness of our results by fitting to different sub-sets of data. We developed our own cosmological computational tools, somewhat different from the standard ones.Comment: 18 pages, 8 figures. Added in v2: an explicit comparison of our code with CAMB, some clarifications on the statistical analysis and some references. Matches version published in JCA

    An optimally concentrated Gabor transform for localized time-frequency components

    Full text link
    Gabor analysis is one of the most common instances of time-frequency signal analysis. Choosing a suitable window for the Gabor transform of a signal is often a challenge for practical applications, in particular in audio signal processing. Many time-frequency (TF) patterns of different shapes may be present in a signal and they can not all be sparsely represented in the same spectrogram. We propose several algorithms, which provide optimal windows for a user-selected TF pattern with respect to different concentration criteria. We base our optimization algorithm on lpl^p-norms as measure of TF spreading. For a given number of sampling points in the TF plane we also propose optimal lattices to be used with the obtained windows. We illustrate the potentiality of the method on selected numerical examples

    MCMC Methods for Multi-Response Generalized Linear Mixed Models: The MCMCglmm R Package

    Get PDF
    Generalized linear mixed models provide a flexible framework for modeling a range of data, although with non-Gaussian response variables the likelihood cannot be obtained in closed form. Markov chain Monte Carlo methods solve this problem by sampling from a series of simpler conditional distributions that can be evaluated. The R package MCMCglmm implements such an algorithm for a range of model fitting problems. More than one response variable can be analyzed simultaneously, and these variables are allowed to follow Gaussian, Poisson, multi(bi)nominal, exponential, zero-inflated and censored distributions. A range of variance structures are permitted for the random effects, including interactions with categorical or continuous variables (i.e., random regression), and more complicated variance structures that arise through shared ancestry, either through a pedigree or through a phylogeny. Missing values are permitted in the response variable(s) and data can be known up to some level of measurement error as in meta-analysis. All simu- lation is done in C/ C++ using the CSparse library for sparse linear systems.

    Conjugate Projective Limits

    Full text link
    We characterize conjugate nonparametric Bayesian models as projective limits of conjugate, finite-dimensional Bayesian models. In particular, we identify a large class of nonparametric models representable as infinite-dimensional analogues of exponential family distributions and their canonical conjugate priors. This class contains most models studied in the literature, including Dirichlet processes and Gaussian process regression models. To derive these results, we introduce a representation of infinite-dimensional Bayesian models by projective limits of regular conditional probabilities. We show under which conditions the nonparametric model itself, its sufficient statistics, and -- if they exist -- conjugate updates of the posterior are projective limits of their respective finite-dimensional counterparts. We illustrate our results both by application to existing nonparametric models and by construction of a model on infinite permutations.Comment: 49 pages; improved version: revised proof of theorem 3 (results unchanged), discussion added, exposition revise

    Max-Sliced Wasserstein Distance and its use for GANs

    Full text link
    Generative adversarial nets (GANs) and variational auto-encoders have significantly improved our distribution modeling capabilities, showing promise for dataset augmentation, image-to-image translation and feature learning. However, to model high-dimensional distributions, sequential training and stacked architectures are common, increasing the number of tunable hyper-parameters as well as the training time. Nonetheless, the sample complexity of the distance metrics remains one of the factors affecting GAN training. We first show that the recently proposed sliced Wasserstein distance has compelling sample complexity properties when compared to the Wasserstein distance. To further improve the sliced Wasserstein distance we then analyze its `projection complexity' and develop the max-sliced Wasserstein distance which enjoys compelling sample complexity while reducing projection complexity, albeit necessitating a max estimation. We finally illustrate that the proposed distance trains GANs on high-dimensional images up to a resolution of 256x256 easily.Comment: Accepted to CVPR 201
    • …
    corecore