87,655 research outputs found
Some sharp performance bounds for least squares regression with regularization
We derive sharp performance bounds for least squares regression with
regularization from parameter estimation accuracy and feature selection quality
perspectives. The main result proved for regularization extends a similar
result in [Ann. Statist. 35 (2007) 2313--2351] for the Dantzig selector. It
gives an affirmative answer to an open question in [Ann. Statist. 35 (2007)
2358--2364]. Moreover, the result leads to an extended view of feature
selection that allows less restrictive conditions than some recent work. Based
on the theoretical insights, a novel two-stage -regularization procedure
with selective penalization is analyzed. It is shown that if the target
parameter vector can be decomposed as the sum of a sparse parameter vector with
large coefficients and another less sparse vector with relatively small
coefficients, then the two-stage procedure can lead to improved performance.Comment: Published in at http://dx.doi.org/10.1214/08-AOS659 the Annals of
Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical
Statistics (http://www.imstat.org
Multi-stage Convex Relaxation for Feature Selection
A number of recent work studied the effectiveness of feature selection using
Lasso. It is known that under the restricted isometry properties (RIP), Lasso
does not generally lead to the exact recovery of the set of nonzero
coefficients, due to the looseness of convex relaxation. This paper considers
the feature selection property of nonconvex regularization, where the solution
is given by a multi-stage convex relaxation scheme. Under appropriate
conditions, we show that the local solution obtained by this procedure recovers
the set of nonzero coefficients without suffering from the bias of Lasso
relaxation, which complements parameter estimation results of this procedure
Sparse Recovery with Orthogonal Matching Pursuit under RIP
This paper presents a new analysis for the orthogonal matching pursuit (OMP)
algorithm. It is shown that if the restricted isometry property (RIP) is
satisfied at sparsity level , then OMP can recover a
-sparse signal in 2-norm. For compressed sensing applications, this
result implies that in order to uniformly recover a -sparse signal in
\Real^d, only random projections are needed. This analysis
improves earlier results on OMP that depend on stronger conditions such as
mutual incoherence that can only be satisfied with
random projections
From -entropy to KL-entropy: Analysis of minimum information complexity density estimation
We consider an extension of -entropy to a KL-divergence based
complexity measure for randomized density estimation methods. Based on this
extension, we develop a general information-theoretical inequality that
measures the statistical complexity of some deterministic and randomized
density estimators. Consequences of the new inequality will be presented. In
particular, we show that this technique can lead to improvements of some
classical results concerning the convergence of minimum description length and
Bayesian posterior distributions. Moreover, we are able to derive clean
finite-sample convergence bounds that are not obtainable using previous
approaches.Comment: Published at http://dx.doi.org/10.1214/009053606000000704 in the
Annals of Statistics (http://www.imstat.org/aos/) by the Institute of
Mathematical Statistics (http://www.imstat.org
A General Framework of Dual Certificate Analysis for Structured Sparse Recovery Problems
This paper develops a general theoretical framework to analyze structured
sparse recovery problems using the notation of dual certificate. Although
certain aspects of the dual certificate idea have already been used in some
previous work, due to the lack of a general and coherent theory, the analysis
has so far only been carried out in limited scopes for specific problems. In
this context the current paper makes two contributions. First, we introduce a
general definition of dual certificate, which we then use to develop a unified
theory of sparse recovery analysis for convex programming. Second, we present a
class of structured sparsity regularization called structured Lasso for which
calculations can be readily performed under our theoretical framework. This new
theory includes many seemingly loosely related previous work as special cases;
it also implies new results that improve existing ones even for standard
formulations such as L1 regularization
- …
