22 research outputs found
Detection of a sparse submatrix of a high-dimensional noisy matrix
We observe a matrix with i.i.d. in , and . We test the
null hypothesis for all against the alternative that there
exists some submatrix of size with significant elements in the
sense that . We propose a test procedure and compute the
asymptotical detection boundary so that the maximal testing risk tends to 0
as , , , . We prove that this
boundary is asymptotically sharp minimax under some additional constraints.
Relations with other testing problems are discussed. We propose a testing
procedure which adapts to unknown within some given set and compute the
adaptive sharp rates. The implementation of our test procedure on synthetic
data shows excellent behavior for sparse, not necessarily squared matrices. We
extend our sharp minimax results in different directions: first, to Gaussian
matrices with unknown variance, next, to matrices of random variables having a
distribution from an exponential family (non-Gaussian) and, finally, to a
two-sided alternative for matrices with Gaussian elements.Comment: Published in at http://dx.doi.org/10.3150/12-BEJ470 the Bernoulli
(http://isi.cbs.nl/bernoulli/) by the International Statistical
Institute/Bernoulli Society (http://isi.cbs.nl/BS/bshome.htm
Statistical inference in compound functional models
We consider a general nonparametric regression model called the compound
model. It includes, as special cases, sparse additive regression and
nonparametric (or linear) regression with many covariates but possibly a small
number of relevant covariates. The compound model is characterized by three
main parameters: the structure parameter describing the "macroscopic" form of
the compound function, the "microscopic" sparsity parameter indicating the
maximal number of relevant covariates in each component and the usual
smoothness parameter corresponding to the complexity of the members of the
compound. We find non-asymptotic minimax rate of convergence of estimators in
such a model as a function of these three parameters. We also show that this
rate can be attained in an adaptive way
Adaptation in minimax nonparametric hypothesis testing for ellipsoids and Besov bodies
We observe an infinitely dimensional Gaussian random vector
where
is a sequence of standard Gaussian variables and is an
unknown mean. Let
be sets
which correspond to -ellipsoids %of the radiuses
of power semi-axes with -ellipsoid
%of the radiuses and
of semi-axes removed or
to similar Besov bodies with Besov
bodies removed. Here
or
are the parameters which define the sets
for given radiuses ,
is
asymptotical parameter.
For the case is known hypothesis testing problem versus
alternatives
have been considered by Ingster and Suslina [11]
in minimax setting.
It was shown that there is a partition of the set of on to regions
with different types of asymptotics: classical, trivial, degenerate and
Gaussian (of two main and some "boundary" types). Also there is essential
dependence of the structure of asymptotically minimax tests on the
parameter for the case of Gaussian asymptotics .
In this paper we consider
alternative for sets
.
This corresponds to adaptive setting: is unknown,
for a compact where and
are regions of main tapes of Gaussian asymptotics .
First the problems of such types were studied by Spokoiny [16, 17].
For ellipsoidal case we study sharp asymptotics of minimax
second kind errors
and construct asymptotically minimax tests.
% .
These asymptotics are analogous to degenerate type.
For Besov bodies case we obtain exact rates
and construct minimax consistent tests.
Analogous exact rates are obtained in a signal detection
problem for continuous variant of white Gaussian noise model: alternatives
correspond to Besov or Sobolev balls with Sobolev or Besov balls removed.
The study is based on results [11] and on
an extension of methods of this paper for degenerate case
Adaptive detection of high-dimensional signal
Let n-dimensional Gaussian random vector x = ξ + v be observed where ξ is a standard n-dimensional Gaussian vector and v ∈ Rn is the unknown mean. In the papers [3,5] there were studied minimax hypothesis testing problems: to test null - hypothesis H0 : v = 0 against two types of alternatives H1 = H1(θn): v ∈ Vn(θn). The first one corresponds to multi-channels signal detection problem for given value b of a signal and number k of channels containing a signal, θn = (b,k). The second one corresponds to lnq-ball of radius R1,n with the lnp-ball of radius R2,n removed, θn = (R1,n, R2,n,p,q) ∈ R4+. It was shown in [3,5] that often there are essential dependences of the structure of asymptotically minimax tests and of the asymptotics of the minimax second kind errors on parameters θn. These imply the problem: to construct adaptive tests having good minimax property for large enough regions Θn of parameters θn. This problem is studied here. We describe the sets Θn such that adaptation is possible without loss of efficiency. For other sets we present wide enough class of asymptotically exact bounds of adaptive efficiency and construct asymptotically minimax test procedures
Sparse classification boundaries
Given a training sample of size from a -dimensional population, we
wish to allocate a new observation to this population or to the
noise. We suppose that the difference between the distribution of the
population and that of the noise is only in a shift, which is a sparse vector.
For the Gaussian noise, fixed sample size , and the dimension that tends
to infinity, we obtain the sharp classification boundary and we propose
classifiers attaining this boundary. We also give extensions of this result to
the case where the sample size depends on and satisfies the condition
, , and to the case of non-Gaussian
noise satisfying the Cram\'er condition
Minimax signal detection in ill-posed inverse problems
Ill-posed inverse problems arise in various scientific fields. We consider
the signal detection problem for mildly, severely and extremely ill-posed
inverse problems with -ellipsoids (bodies), , for Sobolev,
analytic and generalized analytic classes of functions under the Gaussian white
noise model. We study both rate and sharp asymptotics for the error
probabilities in the minimax setup. By construction, the derived tests are,
often, nonadaptive. Minimax rate-optimal adaptive tests of rather simple
structure are also constructed.Comment: Published in at http://dx.doi.org/10.1214/12-AOS1011 the Annals of
Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical
Statistics (http://www.imstat.org
Detection boundary in sparse regression
We study the problem of detection of a p-dimensional sparse vector of
parameters in the linear regression model with Gaussian noise. We establish the
detection boundary, i.e., the necessary and sufficient conditions for the
possibility of successful detection as both the sample size n and the dimension
p tend to the infinity. Testing procedures that achieve this boundary are also
exhibited. Our results encompass the high-dimensional setting (p>> n). The main
message is that, under some conditions, the detection boundary phenomenon that
has been proved for the Gaussian sequence model, extends to high-dimensional
linear regression. Finally, we establish the detection boundaries when the
variance of the noise is unknown. Interestingly, the detection boundaries
sometimes depend on the knowledge of the variance in a high-dimensional
setting