40 research outputs found

    Pivotal estimation in high-dimensional regression via linear programming

    Full text link
    We propose a new method of estimation in high-dimensional linear regression model. It allows for very weak distributional assumptions including heteroscedasticity, and does not require the knowledge of the variance of random errors. The method is based on linear programming only, so that its numerical implementation is faster than for previously known techniques using conic programs, and it allows one to deal with higher dimensional models. We provide upper bounds for estimation and prediction errors of the proposed estimator showing that it achieves the same rate as in the more restrictive situation of fixed design and i.i.d. Gaussian errors with known variance. Following Gautier and Tsybakov (2011), we obtain the results under weaker sensitivity assumptions than the restricted eigenvalue or assimilated conditions

    Regularity Properties for Sparse Regression

    Full text link
    Statistical and machine learning theory has developed several conditions ensuring that popular estimators such as the Lasso or the Dantzig selector perform well in high-dimensional sparse regression, including the restricted eigenvalue, compatibility, and โ„“q\ell_q sensitivity properties. However, some of the central aspects of these conditions are not well understood. For instance, it is unknown if these conditions can be checked efficiently on any given data set. This is problematic, because they are at the core of the theory of sparse regression. Here we provide a rigorous proof that these conditions are NP-hard to check. This shows that the conditions are computationally infeasible to verify, and raises some questions about their practical applications. However, by taking an average-case perspective instead of the worst-case view of NP-hardness, we show that a particular condition, โ„“q\ell_q sensitivity, has certain desirable properties. This condition is weaker and more general than the others. We show that it holds with high probability in models where the parent population is well behaved, and that it is robust to certain data processing steps. These results are desirable, as they provide guidance about when the condition, and more generally the theory of sparse regression, may be relevant in the analysis of high-dimensional correlated observational data.Comment: Manuscript shortened and more motivation added. To appear in Communications in Mathematics and Statistic

    Optimal False Discovery Control of Minimax Estimator

    Full text link
    In the analysis of high dimensional regression models, there are two important objectives: statistical estimation and variable selection. In literature, most works focus on either optimal estimation, e.g., minimax L2L_2 error, or optimal selection behavior, e.g., minimax Hamming loss. However in this study, we investigate the subtle interplay between the estimation accuracy and selection behavior. Our result shows that an estimator's L2L_2 error rate critically depends on its performance of type I error control. Essentially, the minimax convergence rate of false discovery rate over all rate-minimax estimators is a polynomial of the true sparsity ratio. This result helps us to characterize the false positive control of rate-optimal estimators under different sparsity regimes. More specifically, under near-linear sparsity, the number of yielded false positives always explodes to infinity under worst scenario, but the false discovery rate still converges to 0; under linear sparsity, even the false discovery rate doesn't asymptotically converge to 0. On the other side, in order to asymptotically eliminate all false discoveries, the estimator must be sub-optimal in terms of its convergence rate. This work attempts to offer rigorous analysis on the incompatibility phenomenon between selection consistency and rate-minimaxity observed in the high dimensional regression literature

    ์ดˆ๊ณ ์ฐจ์› ์ž๋ฃŒ์— ๋Œ€ํ•œ ๊ต์ • ๋น„๋ณผ๋ก ๋ฒŒ์ ํ™” ๋กœ์ง€์Šคํ‹ฑ ํšŒ๊ท€๋ถ„์„

    Get PDF
    ํ•™์œ„๋…ผ๋ฌธ(๋ฐ•์‚ฌ)--์„œ์šธ๋Œ€ํ•™๊ต ๋Œ€ํ•™์› :์ž์—ฐ๊ณผํ•™๋Œ€ํ•™ ํ†ต๊ณ„ํ•™๊ณผ,2019. 8. ๊น€์šฉ๋Œ€.In high dimensional linear regression, penalized regression methods are used for estimation and variable selection simultaneously. The LASSO is a penalized regression method which is easy to compute the solution, but the LASSO solution is hard to satisfy the variable selection consistency. Nonconvex penalized regression methods such as the SCAD and the MCP have the oracle property which contains variable selection consistency. However, direct computation of the global solution to the nonconvex penalized regression is infeasible. The calibrated CCCP is developed which can obtain the oracle estimator as the unique local minimum. We propose the calibrated CCCP for logistic model. We prove that the calibrated CCCP for logistic model produces a consistent solution path which contains the oracle estimator with probability tending to one. Since the loss function for logistic model is not quadratic, we apply the MLQA-CCCP algorithm for the penalized objective function. Furthermore, we extend the theoretical result to the case of Huber loss instead of the logistic loss. The numerical experiments support our theoretical results.๊ณ ์ฐจ์› ์„ ํ˜•ํšŒ๊ท€๋ถ„์„์—์„œ ๋ฒŒ์ ํ™” ํšŒ๊ท€ ๋ฐฉ๋ฒ•์€ ์ถ”์ •๊ณผ ๋ณ€์ˆ˜์„ ํƒ์„ ๋™์‹œ์— ํ•˜๋Š” ๋ฐฉ๋ฒ•์ด๋‹ค. ๋ผ์†Œ๋Š” ๋ฒŒ์ ํ™” ํšŒ๊ท€ ๋ฐฉ๋ฒ•์˜ ํ•œ๊ฐ€์ง€๋กœ, ๊ทธ ํ•ด๋ฅผ ๊ตฌํ•˜๊ธฐ ์‰ฝ๋‹ค๋Š” ์žฅ์ ์ด ์žˆ์œผ๋‚˜ ๋ณ€์ˆ˜์„ ํƒ ์ผ์น˜์„ฑ์„ ๋งŒ์กฑํ•˜๊ธฐ ์–ด๋ ต๋‹ค. MCP์™€ SCAD ๋“ฑ๊ณผ ๊ฐ™์€ ๋น„๋ณผ๋ก ๋ฒŒ์ ํ™” ํšŒ๊ท€ ๋ฐฉ๋ฒ•์€ ๋ณ€์ˆ˜์„ ํƒ ์ผ์น˜์„ฑ์„ ํฌํ•จํ•œ ์‹ ์˜ ์„ฑ์งˆ์„ ๊ฐ€์ง„๋‹ค. ๊ทธ๋Ÿฌ๋‚˜ ๋น„๋ณผ๋ก ๋ฒŒ์ ํ™” ํšŒ๊ท€์—์„œ ์ „์—ญ ์ตœ์ ํ•ด์˜ ์ง์ ‘์ ์ธ ๊ณ„์‚ฐ์ด ์–ด๋ ค์›Œ ์‹ ์˜ ์ถ”์ •๋Ÿ‰์„ ๊ตฌํ•˜๊ธฐ๊ฐ€ ์–ด๋ ต๋‹ค. ํ•œํŽธ, ์กฐ์ •๋œ CCCP ์•Œ๊ณ ๋ฆฌ์ฆ˜์œผ๋กœ ๊ตฌํ•œ ์œ ์ผํ•œ ๊ตญ์†Œ ์ตœ์†Œํ•ด๋Š” ์‹ ์˜ ์ถ”์ •๋Ÿ‰์ด ๋œ๋‹ค๋Š” ์ด๋ก ์  ์‚ฌ์‹ค์ด ์•Œ๋ ค์ ธ์žˆ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์—์„œ๋Š” ๋กœ์ง€์Šคํ‹ฑ ๋ชจํ˜•์— ๋Œ€ํ•œ ์กฐ์ •๋œ CCCP ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์ œ์•ˆํ•œ๋‹ค. ๊ทธ๋ฆฌ๊ณ  ๋กœ์ง€์Šคํ‹ฑ ๋ชจํ˜•์˜ ์กฐ์ •๋œ CCCP ์•Œ๊ณ ๋ฆฌ์ฆ˜์œผ๋กœ ๊ณ„์‚ฐ๋œ ํ•ด๊ฐ€ 1๋กœ ํ–ฅํ•ด๊ฐ€๋Š” ํ™•๋ฅ ๋กœ ์‹ ์˜ ์ถ”์ •๋Ÿ‰์ด ๋จ์„ ์ฆ๋ช…ํ•œ๋‹ค. ๋กœ์ง€์Šคํ‹ฑ๋ชจํ˜•์—์„œ๋Š” ์†์‹คํ•จ์ˆ˜๊ฐ€ 2์ฐจํ•จ์ˆ˜๊ฐ€ ์•„๋‹ˆ๊ธฐ ๋•Œ๋ฌธ์—, MLQA-CCCP ์•Œ๊ณ ๋ฆฌ์ฆ˜์„ ์ ์šฉํ•˜์˜€๋‹ค. ๋˜ํ•œ, ๋กœ์ง€์Šคํ‹ฑ ์†์‹คํ•จ์ˆ˜๋ฅผ ํ™•์žฅํ•˜์—ฌ Huber ์†์‹คํ•จ์ˆ˜์—์„œ๋„ ๊ฐ™์€ ๊ฒฐ๊ณผ๊ฐ€ ์„ฑ๋ฆฝํ•จ์„ ์ฆ๋ช…ํ•œ๋‹ค. ๋ณธ ๋…ผ๋ฌธ์˜ ์ˆ˜์น˜ ์‹คํ—˜๋“ค์€ ์ด๋ก ์  ๊ฒฐ๊ณผ๋“ค์„ ๋’ท๋ฐ›์นจํ•œ๋‹ค.1. Introduction 1 1.1 Overview 1 1.2 Outline of the thesis 4 2. Literature review : Penalized Regression on High Dimensional Regression 5 2.1 Introduction 5 2.2 LASSO 8 2.3 Nonconvex penalized regression 12 2.4 The calibrated CCCP[Wang et al.,2013] 18 2.5 Review of compatibility condition 19 2.6 Algorithms for l1 penalized regression 24 3. The calibrated CCCP for logistic model 26 3.1 Introduction 26 3.2 The proposed algorithm 27 3.3 Assumptions 31 3.4 Theoretical properties 34 4. Experiments 44 4.1 Simulation studies 44 4.2 Real data analysis 51 5. Conclusion 54 Bibliography 56 Abstract (in Korean) 62Docto

    Geometric Inference for General High-Dimensional Linear Inverse Problems

    Get PDF
    This paper presents a unified geometric framework for the statistical analysis of a general ill-posed linear inverse model which includes as special cases noisy compressed sensing, sign vector recovery, trace regression, orthogonal matrix estimation and noisy matrix completion. We propose computationally feasible convex programs for statistical inference including estimation, confidence intervals and hypothesis testing. A theoretical framework is developed to characterize the local estimation rate of convergence and to provide statistical inference guarantees. Our results are built based on the local conic geometry and duality. The difficulty of statistical inference is captured by the geometric characterization of the local tangent cone through the Gaussian width and Sudakov estimate
    corecore