2,895 research outputs found

    Boosting for high-dimensional linear models

    Full text link
    We prove that boosting with the squared error loss, L2L_2Boosting, is consistent for very high-dimensional linear models, where the number of predictor variables is allowed to grow essentially as fast as OO(exp(sample size)), assuming that the true underlying regression function is sparse in terms of the â„“1\ell_1-norm of the regression coefficients. In the language of signal processing, this means consistency for de-noising using a strongly overcomplete dictionary if the underlying signal is sparse in terms of the â„“1\ell_1-norm. We also propose here an AIC\mathit{AIC}-based method for tuning, namely for choosing the number of boosting iterations. This makes L2L_2Boosting computationally attractive since it is not required to run the algorithm multiple times for cross-validation as commonly used so far. We demonstrate L2L_2Boosting for simulated data, in particular where the predictor dimension is large in comparison to sample size, and for a difficult tumor-classification problem with gene expression microarray data.Comment: Published at http://dx.doi.org/10.1214/009053606000000092 in the Annals of Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical Statistics (http://www.imstat.org

    Statistical significance in high-dimensional linear models

    Full text link
    We propose a method for constructing p-values for general hypotheses in a high-dimensional linear model. The hypotheses can be local for testing a single regression parameter or they may be more global involving several up to all parameters. Furthermore, when considering many hypotheses, we show how to adjust for multiple testing taking dependence among the p-values into account. Our technique is based on Ridge estimation with an additional correction term due to a substantial projection bias in high dimensions. We prove strong error control for our p-values and provide sufficient conditions for detection: for the former, we do not make any assumption on the size of the true underlying regression coefficients while regarding the latter, our procedure might not be optimal in terms of power. We demonstrate the method in simulated examples and a real data application.Comment: Published in at http://dx.doi.org/10.3150/12-BEJSP11 the Bernoulli (http://isi.cbs.nl/bernoulli/) by the International Statistical Institute/Bernoulli Society (http://isi.cbs.nl/BS/bshome.htm

    Introduction to the Lehmann special section

    Full text link
    The current Special Issue of The Annals of Statistics contains three invited articles. Javier Rojo discusses Erich's scientific achievements and provides complete lists of his scientific writings and his former Ph.D. students. Willem van Zwet describes aspects of Erich's life and work, enriched with personal and interesting anecdotes of Erich's long and productive scientific journey. Finally, Peter Bickel, Aiyou Chen and Elizaveta Levina present a research paper on network models: they dedicate their contribution to Erich, emphasizing that their new nonparametric method and issues about optimality have been very much influenced by Erich's thinking.Comment: Published in at http://dx.doi.org/10.1214/11-AOS928 the Annals of Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical Statistics (http://www.imstat.org

    Discussion: One-step sparse estimates in nonconcave penalized likelihood models

    Full text link
    Discussion of ``One-step sparse estimates in nonconcave penalized likelihood models'' [arXiv:0808.1012]Comment: Published in at http://dx.doi.org/10.1214/07-AOS0316A the Annals of Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical Statistics (http://www.imstat.org

    Marginal integration for nonparametric causal inference

    Full text link
    We consider the problem of inferring the total causal effect of a single variable intervention on a (response) variable of interest. We propose a certain marginal integration regression technique for a very general class of potentially nonlinear structural equation models (SEMs) with known structure, or at least known superset of adjustment variables: we call the procedure S-mint regression. We easily derive that it achieves the convergence rate as for nonparametric regression: for example, single variable intervention effects can be estimated with convergence rate n−2/5n^{-2/5} assuming smoothness with twice differentiable functions. Our result can also be seen as a major robustness property with respect to model misspecification which goes much beyond the notion of double robustness. Furthermore, when the structure of the SEM is not known, we can estimate (the equivalence class of) the directed acyclic graph corresponding to the SEM, and then proceed by using S-mint based on these estimates. We empirically compare the S-mint regression method with more classical approaches and argue that the former is indeed more robust, more reliable and substantially simpler.Comment: 40 pages, 14 figure
    • …
    corecore