427 research outputs found

    Direct formulation to Cholesky decomposition of a general nonsingular correlation matrix

    Get PDF
    We present two novel, explicit representations of Cholesky factor of a nonsingular correlation matrix. The first representation uses semi-partial correlation coefficients as its entries. The second, uses an equivalent form of the square roots of the differences between two ratios of successive determinants. Each of the two new forms enjoys parsimony of notations and offers a simpler alternative to both spherical factorization and the multiplicative partial correlation Cholesky matrix (Cooke et al 2011). Two relevant applications are offered for each form: a simple tt-test for assessing the independence of a single variable in a multivariate normal structure, and a straightforward algorithm for generating random positive-definite correlation matrix. The second representation is also extended to any nonsingular hermitian matrix.Comment: Accepted to Statistics and Probability Letters, March 201

    A fast Metropolis-Hastings method for generating random correlation matrices

    Full text link
    We propose a novel Metropolis-Hastings algorithm to sample uniformly from the space of correlation matrices. Existing methods in the literature are based on elaborated representations of a correlation matrix, or on complex parametrizations of it. By contrast, our method is intuitive and simple, based the classical Cholesky factorization of a positive definite matrix and Markov chain Monte Carlo theory. We perform a detailed convergence analysis of the resulting Markov chain, and show how it benefits from fast convergence, both theoretically and empirically. Furthermore, in numerical experiments our algorithm is shown to be significantly faster than the current alternative approaches, thanks to its simple yet principled approach.Comment: 8 pages, 3 figures, 2018 conferenc

    Symmetric matrices, Catalan paths, and correlations

    Get PDF
    Kenyon and Pemantle (2014) gave a formula for the entries of a square matrix in terms of connected principal and almost-principal minors. Each entry is an explicit Laurent polynomial whose terms are the weights of domino tilings of a half Aztec diamond. They conjectured an analogue of this parametrization for symmetric matrices, where the Laurent monomials are indexed by Catalan paths. In this paper we prove the Kenyon-Pemantle conjecture, and apply this to a statistics problem pioneered by Joe (2006). Correlation matrices are represented by an explicit bijection from the cube to the elliptope

    Parsimony in model selection: tools for assessing fit propensity

    Get PDF
    Theories can be represented as statistical models for empirical testing. There is a vast literature on model selection and multimodel inference that focuses on how to assess which statistical model, and therefore which theory, best fits the available data. For example, given some data, one can compare models on various information criterion or other fit statistics. However, what these indices fail to capture is the full range of counterfactuals. That is, some models may fit the given data better not because they represent a more correct theory, but simply because these models have more fit propensity - a tendency to fit a wider range of data, even nonsensical data, better. Current approaches fall short in considering the principle of parsimony (Occam’s Razor), often equating it with the number of model parameters. Here we offer a toolkit for researchers to better study and understand parsimony through the fit propensity of Structural Equation Models. We provide an R package (ockhamSEM) built on the popular lavaan package. To illustrate the importance of evaluating fit propensity, we use ockhamSEM to investigate the factor structure of the Rosenberg Self-Esteem Scale

    A partial correlation vine based approach for modeling and forecasting multivariate volatility time-series

    Full text link
    A novel approach for dynamic modeling and forecasting of realized covariance matrices is proposed. Realized variances and realized correlation matrices are jointly estimated. The one-to-one relationship between a positive definite correlation matrix and its associated set of partial correlations corresponding to any vine specification is used for data transformation. The model components therefore are realized variances as well as realized standard and partial correlations corresponding to a daily log-return series. As such, they have a clear practical interpretation. A method to select a regular vine structure, which allows for parsimonious time-series and dependence modeling of the model components, is introduced. Being algebraically independent the latter do not underlie any algebraic constraint. The proposed model approach is outlined in detail and motivated along with a real data example on six highly liquid stocks. The forecasting performance is evaluated both with respect to statistical precision and in the context of portfolio optimization. Comparisons with Cholesky decomposition based benchmark models support the excellent prediction ability of the proposed model approach

    A method for generating realistic correlation matrices

    Get PDF
    Simulating sample correlation matrices is important in many areas of statistics. Approaches such as generating Gaussian data and finding their sample correlation matrix or generating random uniform [−1,1][-1,1] deviates as pairwise correlations both have drawbacks. We develop an algorithm for adding noise, in a highly controlled manner, to general correlation matrices. In many instances, our method yields results which are superior to those obtained by simply simulating Gaussian data. Moreover, we demonstrate how our general algorithm can be tailored to a number of different correlation models. Using our results with a few different applications, we show that simulating correlation matrices can help assess statistical methodology.Comment: Published in at http://dx.doi.org/10.1214/13-AOAS638 the Annals of Applied Statistics (http://www.imstat.org/aoas/) by the Institute of Mathematical Statistics (http://www.imstat.org

    Comparison of goodness measures for linear factor structures

    Get PDF
    Linear factor structures often exist in empirical data, and they can be mapped by factor analysis. It is, however, not straightforward how to measure the goodness of a factor analysis solution since its results should correspond to various requirements. Instead of a unique indicator, several goodness measures can be defined that all contribute to the evaluation of the results. This paper aims to find an answer to the question whether factor analysis outputs can meet several goodness criteria at the same time. Data aggregability (measured by the determinant of the correlation matrix and the proportion of explained variance) and the extent of latency (defined by the determinant of the antiimage correlation matrix, the maximum partial correlation coefficient and the Kaiser–Meyer–Olkin measure of sampling adequacy) are studied. According to the theoretical and simulation results, it is not possible to meet simultaneously these two criteria when the correlation matrices are relatively small. For larger correlation matrices, however, there are linear factor structures that combine good data aggregability with a high extent of latency
    • …
    corecore