12,694 research outputs found

    Bayesian dimensionality reduction with PCA using penalized semi-integrated likelihood

    Full text link
    We discuss the problem of estimating the number of principal components in Principal Com- ponents Analysis (PCA). Despite of the importance of the problem and the multitude of solutions proposed in the literature, it comes as a surprise that there does not exist a coherent asymptotic framework which would justify different approaches depending on the actual size of the data set. In this paper we address this issue by presenting an approximate Bayesian approach based on Laplace approximation and introducing a general method for building the model selection criteria, called PEnalized SEmi-integrated Likelihood (PESEL). Our general framework encompasses a variety of existing approaches based on probabilistic models, like e.g. Bayesian Information Criterion for the Probabilistic PCA (PPCA), and allows for construction of new criteria, depending on the size of the data set at hand. Specifically, we define PESEL when the number of variables substantially exceeds the number of observations. We also report results of extensive simulation studies and real data analysis, which illustrate good properties of our proposed criteria as compared to the state-of- the-art methods and very recent proposals. Specifially, these simulations show that PESEL based criteria can be quite robust against deviations from the probabilistic model assumptions. Selected PESEL based criteria for the estimation of the number of principal components are implemented in R package varclust, which is available on github (https://github.com/psobczyk/varclust).Comment: 31 pages, 7 figure

    Decomposition of multicomponent mass spectra using Bayesian probability theory

    Full text link
    We present a method for the decomposition of mass spectra of mixture gases using Bayesian probability theory. The method works without any calibration measurement and therefore applies also to the analysis of spectra containing unstable species. For the example of mixtures of three different hydrocarbon gases the algorithm provides concentrations and cracking coefficients of each mixture component as well as their confidence intervals. The amount of information needed to obtain reliable results and its relation to the accuracy of our analysis are discussed

    Bayesian Identification of Elastic Constants in Multi-Directional Laminate from Moir\'e Interferometry Displacement Fields

    Get PDF
    The ply elastic constants needed for classical lamination theory analysis of multi-directional laminates may differ from those obtained from unidirectional laminates because of three dimensional effects. In addition, the unidirectional laminates may not be available for testing. In such cases, full-field displacement measurements offer the potential of identifying several material properties simultaneously. For that, it is desirable to create complex displacement fields that are strongly influenced by all the elastic constants. In this work, we explore the potential of using a laminated plate with an open-hole under traction loading to achieve that and identify all four ply elastic constants (E 1, E 2, 12, G 12) at once. However, the accuracy of the identified properties may not be as good as properties measured from individual tests due to the complexity of the experiment, the relative insensitivity of the measured quantities to some of the properties and the various possible sources of uncertainty. It is thus important to quantify the uncertainty (or confidence) with which these properties are identified. Here, Bayesian identification is used for this purpose, because it can readily model all the uncertainties in the analysis and measurements, and because it provides the full coupled probability distribution of the identified material properties. In addition, it offers the potential to combine properties identified based on substantially different experiments. The full-field measurement is obtained by moir\'e interferometry. For computational efficiency the Bayesian approach was applied to a proper orthogonal decomposition (POD) of the displacement fields. The analysis showed that the four orthotropic elastic constants are determined with quite different confidence levels as well as with significant correlation. Comparison with manufacturing specifications showed substantial difference in one constant, and this conclusion agreed with earlier measurement of that constant by a traditional four-point bending test. It is possible that the POD approach did not take full advantage of the copious data provided by the full field measurements, and for that reason that data is provided for others to use (as on line material attached to the article)

    Bayesian Inference on Matrix Manifolds for Linear Dimensionality Reduction

    Full text link
    We reframe linear dimensionality reduction as a problem of Bayesian inference on matrix manifolds. This natural paradigm extends the Bayesian framework to dimensionality reduction tasks in higher dimensions with simpler models at greater speeds. Here an orthogonal basis is treated as a single point on a manifold and is associated with a linear subspace on which observations vary maximally. Throughout this paper, we employ the Grassmann and Stiefel manifolds for various dimensionality reduction problems, explore the connection between the two manifolds, and use Hybrid Monte Carlo for posterior sampling on the Grassmannian for the first time. We delineate in which situations either manifold should be considered. Further, matrix manifold models are used to yield scientific insight in the context of cognitive neuroscience, and we conclude that our methods are suitable for basic inference as well as accurate prediction.Comment: All datasets and computer programs are publicly available at http://www.ics.uci.edu/~babaks/Site/Codes.htm

    Efficient and automatic methods for flexible regression on spatiotemporal data, with applications to groundwater monitoring

    Get PDF
    Fitting statistical models to spatiotemporal data requires finding the right balance between imposing smoothness and following the data. In the context of P-splines, we propose a Bayesian framework for choosing the smoothing parameter which allows the construction of fully-automatic data-driven methods for fitting flexible models to spatiotemporal data. An implementation, which is highly computationally efficient and which exploits the sparsity of the design and penalty matrices, is proposed. The findings are illustrated using a simulation study and two examples, all concerned with the modelling of contaminants in groundwater. This suggests that the proposed strategy is more stable that competing methods based on the use of criteria such as GCV and AIC

    Bayesian inference in the time varying cointegration model

    Get PDF
    There are both theoretical and empirical reasons for believing that the parameters of macroeconomic models may vary over time. However, work with time-varying parameter models has largely involved Vector autoregressions (VARs), ignoring cointegration. This is despite the fact that cointegration plays an important role in informing macroeconomists on a range of issues. In this paper we develop time varying parameter models which permit coin- tegration. Time-varying parameter VARs (TVP-VARs) typically use state space representations to model the evolution of parameters. In this paper, we show that it is not sensible to use straightforward extensions of TVP-VARs when allowing for cointegration. Instead we develop a speci…cation which allows for the cointegrating space to evolve over time in a manner comparable to the random walk variation used with TVP-VARs. The properties of our approach are investigated before developing a method of posterior simulation. We use our methods in an empirical investigation involving a permanent/transitory variance decomposition for inflation
    • …
    corecore