40 research outputs found
Pivotal estimation in high-dimensional regression via linear programming
We propose a new method of estimation in high-dimensional linear regression
model. It allows for very weak distributional assumptions including
heteroscedasticity, and does not require the knowledge of the variance of
random errors. The method is based on linear programming only, so that its
numerical implementation is faster than for previously known techniques using
conic programs, and it allows one to deal with higher dimensional models. We
provide upper bounds for estimation and prediction errors of the proposed
estimator showing that it achieves the same rate as in the more restrictive
situation of fixed design and i.i.d. Gaussian errors with known variance.
Following Gautier and Tsybakov (2011), we obtain the results under weaker
sensitivity assumptions than the restricted eigenvalue or assimilated
conditions
Regularity Properties for Sparse Regression
Statistical and machine learning theory has developed several conditions
ensuring that popular estimators such as the Lasso or the Dantzig selector
perform well in high-dimensional sparse regression, including the restricted
eigenvalue, compatibility, and sensitivity properties. However, some
of the central aspects of these conditions are not well understood. For
instance, it is unknown if these conditions can be checked efficiently on any
given data set. This is problematic, because they are at the core of the theory
of sparse regression.
Here we provide a rigorous proof that these conditions are NP-hard to check.
This shows that the conditions are computationally infeasible to verify, and
raises some questions about their practical applications.
However, by taking an average-case perspective instead of the worst-case view
of NP-hardness, we show that a particular condition, sensitivity, has
certain desirable properties. This condition is weaker and more general than
the others. We show that it holds with high probability in models where the
parent population is well behaved, and that it is robust to certain data
processing steps. These results are desirable, as they provide guidance about
when the condition, and more generally the theory of sparse regression, may be
relevant in the analysis of high-dimensional correlated observational data.Comment: Manuscript shortened and more motivation added. To appear in
Communications in Mathematics and Statistic
Optimal False Discovery Control of Minimax Estimator
In the analysis of high dimensional regression models, there are two
important objectives: statistical estimation and variable selection. In
literature, most works focus on either optimal estimation, e.g., minimax
error, or optimal selection behavior, e.g., minimax Hamming loss. However in
this study, we investigate the subtle interplay between the estimation accuracy
and selection behavior. Our result shows that an estimator's error rate
critically depends on its performance of type I error control. Essentially, the
minimax convergence rate of false discovery rate over all rate-minimax
estimators is a polynomial of the true sparsity ratio. This result helps us to
characterize the false positive control of rate-optimal estimators under
different sparsity regimes. More specifically, under near-linear sparsity, the
number of yielded false positives always explodes to infinity under worst
scenario, but the false discovery rate still converges to 0; under linear
sparsity, even the false discovery rate doesn't asymptotically converge to 0.
On the other side, in order to asymptotically eliminate all false discoveries,
the estimator must be sub-optimal in terms of its convergence rate. This work
attempts to offer rigorous analysis on the incompatibility phenomenon between
selection consistency and rate-minimaxity observed in the high dimensional
regression literature
์ด๊ณ ์ฐจ์ ์๋ฃ์ ๋ํ ๊ต์ ๋น๋ณผ๋ก ๋ฒ์ ํ ๋ก์ง์คํฑ ํ๊ท๋ถ์
ํ์๋
ผ๋ฌธ(๋ฐ์ฌ)--์์ธ๋ํ๊ต ๋ํ์ :์์ฐ๊ณผํ๋ํ ํต๊ณํ๊ณผ,2019. 8. ๊น์ฉ๋.In high dimensional linear regression, penalized regression methods are used for estimation and variable selection simultaneously. The LASSO is a penalized regression method which is easy to compute the solution, but the LASSO solution is hard to satisfy the variable selection consistency. Nonconvex penalized regression methods such as the SCAD and the MCP have the oracle property which contains variable selection consistency. However, direct computation of the global solution to the nonconvex penalized regression is infeasible. The calibrated CCCP is developed which can obtain the oracle estimator as the unique local minimum.
We propose the calibrated CCCP for logistic model. We prove that the calibrated CCCP for logistic model produces a consistent solution path which contains the oracle estimator with probability tending to one. Since the loss function for logistic model is not quadratic, we apply the MLQA-CCCP algorithm for the penalized objective function. Furthermore, we extend the theoretical result to the case of Huber loss instead of the logistic loss. The numerical experiments support our theoretical results.๊ณ ์ฐจ์ ์ ํํ๊ท๋ถ์์์ ๋ฒ์ ํ ํ๊ท ๋ฐฉ๋ฒ์ ์ถ์ ๊ณผ ๋ณ์์ ํ์ ๋์์ ํ๋ ๋ฐฉ๋ฒ์ด๋ค. ๋ผ์๋ ๋ฒ์ ํ ํ๊ท ๋ฐฉ๋ฒ์ ํ๊ฐ์ง๋ก, ๊ทธ ํด๋ฅผ ๊ตฌํ๊ธฐ ์ฝ๋ค๋ ์ฅ์ ์ด ์์ผ๋ ๋ณ์์ ํ ์ผ์น์ฑ์ ๋ง์กฑํ๊ธฐ ์ด๋ ต๋ค. MCP์ SCAD ๋ฑ๊ณผ ๊ฐ์ ๋น๋ณผ๋ก ๋ฒ์ ํ ํ๊ท ๋ฐฉ๋ฒ์ ๋ณ์์ ํ ์ผ์น์ฑ์ ํฌํจํ ์ ์ ์ฑ์ง์ ๊ฐ์ง๋ค. ๊ทธ๋ฌ๋ ๋น๋ณผ๋ก ๋ฒ์ ํ ํ๊ท์์ ์ ์ญ ์ต์ ํด์ ์ง์ ์ ์ธ ๊ณ์ฐ์ด ์ด๋ ค์ ์ ์ ์ถ์ ๋์ ๊ตฌํ๊ธฐ๊ฐ ์ด๋ ต๋ค. ํํธ, ์กฐ์ ๋ CCCP ์๊ณ ๋ฆฌ์ฆ์ผ๋ก ๊ตฌํ ์ ์ผํ ๊ตญ์ ์ต์ํด๋ ์ ์ ์ถ์ ๋์ด ๋๋ค๋ ์ด๋ก ์ ์ฌ์ค์ด ์๋ ค์ ธ์๋ค. ๋ณธ ๋
ผ๋ฌธ์์๋ ๋ก์ง์คํฑ ๋ชจํ์ ๋ํ ์กฐ์ ๋ CCCP ์๊ณ ๋ฆฌ์ฆ์ ์ ์ํ๋ค. ๊ทธ๋ฆฌ๊ณ ๋ก์ง์คํฑ ๋ชจํ์ ์กฐ์ ๋ CCCP ์๊ณ ๋ฆฌ์ฆ์ผ๋ก ๊ณ์ฐ๋ ํด๊ฐ 1๋ก ํฅํด๊ฐ๋ ํ๋ฅ ๋ก ์ ์ ์ถ์ ๋์ด ๋จ์ ์ฆ๋ช
ํ๋ค. ๋ก์ง์คํฑ๋ชจํ์์๋ ์์คํจ์๊ฐ 2์ฐจํจ์๊ฐ ์๋๊ธฐ ๋๋ฌธ์, MLQA-CCCP ์๊ณ ๋ฆฌ์ฆ์ ์ ์ฉํ์๋ค. ๋ํ, ๋ก์ง์คํฑ ์์คํจ์๋ฅผ ํ์ฅํ์ฌ Huber ์์คํจ์์์๋ ๊ฐ์ ๊ฒฐ๊ณผ๊ฐ ์ฑ๋ฆฝํจ์ ์ฆ๋ช
ํ๋ค. ๋ณธ ๋
ผ๋ฌธ์ ์์น ์คํ๋ค์ ์ด๋ก ์ ๊ฒฐ๊ณผ๋ค์ ๋ท๋ฐ์นจํ๋ค.1. Introduction 1
1.1 Overview 1
1.2 Outline of the thesis 4
2. Literature review : Penalized Regression on High Dimensional Regression 5
2.1 Introduction 5
2.2 LASSO 8
2.3 Nonconvex penalized regression 12
2.4 The calibrated CCCP[Wang et al.,2013] 18
2.5 Review of compatibility condition 19
2.6 Algorithms for l1 penalized regression 24
3. The calibrated CCCP for logistic model 26
3.1 Introduction 26
3.2 The proposed algorithm 27
3.3 Assumptions 31
3.4 Theoretical properties 34
4. Experiments 44
4.1 Simulation studies 44
4.2 Real data analysis 51
5. Conclusion 54
Bibliography 56
Abstract (in Korean) 62Docto
Geometric Inference for General High-Dimensional Linear Inverse Problems
This paper presents a unified geometric framework for the statistical analysis of a general ill-posed linear inverse model which includes as special cases noisy compressed sensing, sign vector recovery, trace regression, orthogonal matrix estimation and noisy matrix completion. We propose computationally feasible convex programs for statistical inference including estimation, confidence intervals and hypothesis testing. A theoretical framework is developed to characterize the local estimation rate of convergence and to provide statistical inference guarantees. Our results are built based on the local conic geometry and duality. The difficulty of statistical inference is captured by the geometric characterization of the local tangent cone through the Gaussian width and Sudakov estimate