4 research outputs found
Testing the order of a model
This paper deals with order identification for nested models in the i.i.d.
framework. We study the asymptotic efficiency of two generalized likelihood
ratio tests of the order. They are based on two estimators which are proved to
be strongly consistent. A version of Stein's lemma yields an optimal
underestimation error exponent. The lemma also implies that the overestimation
error exponent is necessarily trivial. Our tests admit nontrivial
underestimation error exponents. The optimal underestimation error exponent is
achieved in some situations. The overestimation error can decay exponentially
with respect to a positive power of the number of observations. These results
are proved under mild assumptions by relating the underestimation (resp.
overestimation) error to large (resp. moderate) deviations of the
log-likelihood process. In particular, it is not necessary that the classical
Cram\'{e}r condition be satisfied; namely, the -densities are not
required to admit every exponential moment. Three benchmark examples with
specific difficulties (location mixture of normal distributions, abrupt changes
and various regressions) are detailed so as to illustrate the generality of our
results.Comment: Published at http://dx.doi.org/10.1214/009053606000000344 in the
Annals of Statistics (http://www.imstat.org/aos/) by the Institute of
Mathematical Statistics (http://www.imstat.org
Large-deviation analysis and applications Of learning tree-structured graphical models
Thesis (Ph. D.)--Massachusetts Institute of Technology, Dept. of Electrical Engineering and Computer Science, 2011.This electronic version was submitted by the student author. The certified thesis is available in the Institute Archives and Special Collections.Cataloged from student submitted PDF version of thesis.Includes bibliographical references (p. 213-228).The design and analysis of complexity-reduced representations for multivariate data is important in many scientific and engineering domains. This thesis explores such representations from two different perspectives: deriving and analyzing performance measures for learning tree-structured graphical models and salient feature subset selection for discrimination. Graphical models have proven to be a flexible class of probabilistic models for approximating high-dimensional data. Learning the structure of such models from data is an important generic task. It is known that if the data are drawn from tree-structured distributions, then the algorithm of Chow and Liu (1968) provides an efficient algorithm for finding the tree that maximizes the likelihood of the data. We leverage this algorithm and the theory of large deviations to derive the error exponent of structure learning for discrete and Gaussian graphical models. We determine the extremal tree structures for learning, that is, the structures that lead to the highest and lowest exponents. We prove that the star minimizes the exponent and the chain maximizes the exponent, which means that among all unlabeled trees, the star and the chain are the worst and best for learning respectively. The analysis is also extended to learning foreststructured graphical models by augmenting the Chow-Liu algorithm with a thresholding procedure. We prove scaling laws on the number of samples and the number variables for structure learning to remain consistent in high-dimensions. The next part of the thesis is concerned with discrimination. We design computationally efficient tree-based algorithms to learn pairs of distributions that are specifically adapted to the task of discrimination and show that they perform well on various datasets vis-`a-vis existing tree-based algorithms. We define the notion of a salient set for discrimination using information-theoretic quantities and derive scaling laws on the number of samples so that the salient set can be recovered asymptotically.by Vincent Yan Fu Tan.Ph.D