We consider the problem of robustly testing the norm of a high-dimensional
sparse signal vector under two different observation models. In the first
model, we are given n i.i.d. samples from the distribution
N(θ,Id) (with unknown θ), of which a small
fraction has been arbitrarily corrupted. Under the promise that
∥θ∥0≤s, we want to correctly distinguish whether ∥θ∥2=0
or ∥θ∥2>γ, for some input parameter γ>0. We show that any
algorithm for this task requires n=Ω(slogsed)
samples, which is tight up to logarithmic factors. We also extend our results
to other common notions of sparsity, namely, ∥θ∥q≤s for any 0<q<2. In the second observation model that we consider, the data is generated
according to a sparse linear regression model, where the covariates are i.i.d.
Gaussian and the regression coefficient (signal) is known to be s-sparse.
Here too we assume that an ϵ-fraction of the data is arbitrarily
corrupted. We show that any algorithm that reliably tests the norm of the
regression coefficient requires at least n=Ω(min(slogd,1/γ4)) samples. Our results show that the complexity of
testing in these two settings significantly increases under robustness
constraints. This is in line with the recent observations made in robust mean
testing and robust covariance testing.Comment: Fixed typos, added a figure and discussion sectio