24,230 research outputs found

    Classification with Minimax Fast Rates for Classes of Bayes Rules with Sparse Representation

    Get PDF
    We construct a classifier which attains the rate of convergence logn/n\log n/n under sparsity and margin assumptions. An approach close to the one met in approximation theory for the estimation of function is used to obtain this result. The idea is to develop the Bayes rule in a fundamental system of L2([0,1]d)L^2([0,1]^d) made of indicator of dyadic sets and to assume that coefficients, equal to 1,0or1-1,0 {or} 1, belong to a kind of L1L^1-ball. This assumption can be seen as a sparsity assumption, in the sense that the proportion of coefficients non equal to zero decreases as "frequency" grows. Finally, rates of convergence are obtained by using an usual trade-off between a bias term and a variance term

    Generalized Permutohedra from Probabilistic Graphical Models

    Get PDF
    A graphical model encodes conditional independence relations via the Markov properties. For an undirected graph these conditional independence relations can be represented by a simple polytope known as the graph associahedron, which can be constructed as a Minkowski sum of standard simplices. There is an analogous polytope for conditional independence relations coming from a regular Gaussian model, and it can be defined using multiinformation or relative entropy. For directed acyclic graphical models and also for mixed graphical models containing undirected, directed and bidirected edges, we give a construction of this polytope, up to equivalence of normal fans, as a Minkowski sum of matroid polytopes. Finally, we apply this geometric insight to construct a new ordering-based search algorithm for causal inference via directed acyclic graphical models.Comment: Appendix B is expanded. Final version to appear in SIAM J. Discrete Mat

    Adaptive Bernstein-von Mises theorems in Gaussian white noise

    Get PDF
    We investigate Bernstein-von Mises theorems for adaptive nonparametric Bayesian procedures in the canonical Gaussian white noise model. We consider both a Hilbert space and multiscale setting with applications in L2L^2 and LL^\infty respectively. This provides a theoretical justification for plug-in procedures, for example the use of certain credible sets for sufficiently smooth linear functionals. We use this general approach to construct optimal frequentist confidence sets based on the posterior distribution. We also provide simulations to numerically illustrate our approach and obtain a visual representation of the geometries involved.Comment: 48 pages, 5 figure

    About the non-asymptotic behaviour of Bayes estimators

    Full text link
    This paper investigates the {\em nonasymptotic} properties of Bayes procedures for estimating an unknown distribution from nn i.i.d.\ observations. We assume that the prior is supported by a model (\scr{S},h) (where hh denotes the Hellinger distance) with suitable metric properties involving the number of small balls that are needed to cover larger ones. We also require that the prior put enough probability on small balls. We consider two different situations. The simplest case is the one of a parametric model containing the target density for which we show that the posterior concentrates around the true distribution at rate 1/n1/\sqrt{n}. In the general situation, we relax the parametric assumption and take into account a possible mispecification of the model. Provided that the Kullback-Leibler Information between the true distribution and \scr{S} is finite, we establish risk bounds for the Bayes estimators.Comment: Extended version of a talk given in June 2013 at BNP9 Conference in Amsterdam - 17 page
    corecore