72,025 research outputs found
Group Lasso for high dimensional sparse quantile regression models
This paper studies the statistical properties of the group Lasso estimator
for high dimensional sparse quantile regression models where the number of
explanatory variables (or the number of groups of explanatory variables) is
possibly much larger than the sample size while the number of variables in
"active" groups is sufficiently small. We establish a non-asymptotic bound on
the -estimation error of the estimator. This bound explains
situations under which the group Lasso estimator is potentially
superior/inferior to the -penalized quantile regression estimator in
terms of the estimation error. We also propose a data-dependent choice of the
tuning parameter to make the method more practical, by extending the original
proposal of Belloni and Chernozhukov (2011) for the -penalized
quantile regression estimator. As an application, we analyze high dimensional
additive quantile regression models. We show that under a set of suitable
regularity conditions, the group Lasso estimator can attain the convergence
rate arbitrarily close to the oracle rate. Finally, we conduct simulations
experiments to examine our theoretical results.Comment: 37 pages. Some errors are correcte
Estimation of a sparse group of sparse vectors
We consider a problem of estimating a sparse group of sparse normal mean
vectors. The proposed approach is based on penalized likelihood estimation with
complexity penalties on the number of nonzero mean vectors and the numbers of
their "significant" components, and can be performed by a computationally fast
algorithm. The resulting estimators are developed within Bayesian framework and
can be viewed as MAP estimators. We establish their adaptive minimaxity over a
wide range of sparse and dense settings. The presented short simulation study
demonstrates the efficiency of the proposed approach that successfully competes
with the recently developed sparse group lasso estimator
Nonparametric Independence Screening in Sparse Ultra-High Dimensional Additive Models
A variable screening procedure via correlation learning was proposed Fan and
Lv (2008) to reduce dimensionality in sparse ultra-high dimensional models.
Even when the true model is linear, the marginal regression can be highly
nonlinear. To address this issue, we further extend the correlation learning to
marginal nonparametric learning. Our nonparametric independence screening is
called NIS, a specific member of the sure independence screening. Several
closely related variable screening procedures are proposed. Under the
nonparametric additive models, it is shown that under some mild technical
conditions, the proposed independence screening methods enjoy a sure screening
property. The extent to which the dimensionality can be reduced by independence
screening is also explicitly quantified. As a methodological extension, an
iterative nonparametric independence screening (INIS) is also proposed to
enhance the finite sample performance for fitting sparse additive models. The
simulation results and a real data analysis demonstrate that the proposed
procedure works well with moderate sample size and large dimension and performs
better than competing methods.Comment: 48 page
Sparse additive regression on a regular lattice
We consider estimation in a sparse additive regression model with the design
points on a regular lattice. We establish the minimax convergence rates over
Sobolev classes and propose a Fourier-based rate-optimal estimator which is
adaptive to the unknown sparsity and smoothness of the response function. The
estimator is derived within Bayesian formalism but can be naturally viewed as a
penalized maximum likelihood estimator with the complexity penalties on the
number of nonzero univariate additive components of the response and on the
numbers of the nonzero coefficients of their Fourer expansions. We compare it
with several existing counterparts and perform a short simulation study to
demonstrate its performance
Functional Regression
Functional data analysis (FDA) involves the analysis of data whose ideal
units of observation are functions defined on some continuous domain, and the
observed data consist of a sample of functions taken from some population,
sampled on a discrete grid. Ramsay and Silverman's 1997 textbook sparked the
development of this field, which has accelerated in the past 10 years to become
one of the fastest growing areas of statistics, fueled by the growing number of
applications yielding this type of data. One unique characteristic of FDA is
the need to combine information both across and within functions, which Ramsay
and Silverman called replication and regularization, respectively. This article
will focus on functional regression, the area of FDA that has received the most
attention in applications and methodological development. First will be an
introduction to basis functions, key building blocks for regularization in
functional regression methods, followed by an overview of functional regression
methods, split into three types: [1] functional predictor regression
(scalar-on-function), [2] functional response regression (function-on-scalar)
and [3] function-on-function regression. For each, the role of replication and
regularization will be discussed and the methodological development described
in a roughly chronological manner, at times deviating from the historical
timeline to group together similar methods. The primary focus is on modeling
and methodology, highlighting the modeling structures that have been developed
and the various regularization approaches employed. At the end is a brief
discussion describing potential areas of future development in this field
- …
