2,508 research outputs found

    A tutorial on estimator averaging in spatial point process models

    Get PDF
    Assume that several competing methods are available to estimate a parameter in a given statistical model. The aim of estimator averaging is to provide a new estimator, built as a linear combination of the initial estimators, that achieves better properties, under the quadratic loss, than each individual initial estimator. This contribution provides an accessible and clear overview of the method, and investigates its performances on standard spatial point process models. It is demonstrated that the average estimator clearly improves on standard procedures for the considered models. For each example, the code to implement the method with the R software (which only consists of few lines) is provided

    On a symbolic representation of non-central Wishart random matrices with applications

    Full text link
    By using a symbolic method, known in the literature as the classical umbral calculus, the trace of a non-central Wishart random matrix is represented as the convolution of the trace of its central component and of a formal variable involving traces of its non-centrality matrix. Thanks to this representation, the moments of this random matrix are proved to be a Sheffer polynomial sequence, allowing us to recover several properties. The multivariate symbolic method generalizes the employment of Sheffer representation and a closed form formula for computing joint moments and cumulants (also normalized) is given. By using this closed form formula and a combinatorial device, known in the literature as necklace, an efficient algorithm for their computations is set up. Applications are given to the computation of permanents as well as to the characterization of inherited estimators of cumulants, which turn useful in dealing with minors of non-central Wishart random matrices. An asymptotic approximation of generalized moments involving free probability is proposed.Comment: Journal of Multivariate Analysis (2014

    Algorithms and Hardness for Robust Subspace Recovery

    Full text link
    We consider a fundamental problem in unsupervised learning called \emph{subspace recovery}: given a collection of mm points in Rn\mathbb{R}^n, if many but not necessarily all of these points are contained in a dd-dimensional subspace TT can we find it? The points contained in TT are called {\em inliers} and the remaining points are {\em outliers}. This problem has received considerable attention in computer science and in statistics. Yet efficient algorithms from computer science are not robust to {\em adversarial} outliers, and the estimators from robust statistics are hard to compute in high dimensions. Are there algorithms for subspace recovery that are both robust to outliers and efficient? We give an algorithm that finds TT when it contains more than a dn\frac{d}{n} fraction of the points. Hence, for say d=n/2d = n/2 this estimator is both easy to compute and well-behaved when there are a constant fraction of outliers. We prove that it is Small Set Expansion hard to find TT when the fraction of errors is any larger, thus giving evidence that our estimator is an {\em optimal} compromise between efficiency and robustness. As it turns out, this basic problem has a surprising number of connections to other areas including small set expansion, matroid theory and functional analysis that we make use of here.Comment: Appeared in Proceedings of COLT 201
    • …
    corecore