9 research outputs found
An Adaptive Sublinear-Time Block Sparse Fourier Transform
The problem of approximately computing the dominant Fourier coefficients of a vector quickly, and using few samples in time domain, is known as the Sparse Fourier Transform (sparse FFT) problem. A long line of work on the sparse FFT has resulted in algorithms with runtime [Hassanieh \emph{et al.}, STOC'12] and sample complexity [Indyk \emph{et al.}, FOCS'14]. These results are proved using non-adaptive algorithms, and the latter sample complexity result is essentially the best possible under the sparsity assumption alone: It is known that even adaptive algorithms must use samples [Hassanieh \emph{et al.}, STOC'12]. By {\em adaptive}, we mean being able to exploit previous samples in guiding the selection of further samples. This paper revisits the sparse FFT problem with the added twist that the sparse coefficients approximately obey a -block sparse model. In this model, signal frequencies are clustered in intervals with width in Fourier space, and is the total sparsity. Signals arising in applications are often well approximated by this model with . Our main result is the first sparse FFT algorithm for -block sparse signals with a sample complexity of at constant signal-to-noise ratios, and sublinear runtime. A similar sample complexity was previously achieved in the works on {\em model-based compressive sensing} using random Gaussian measurements, but used runtime. To the best of our knowledge, our result is the first sublinear-time algorithm for model based compressed sensing, and the first sparse FFT result that goes below the sample complexity bound. Interestingly, the aforementioned model-based compressive sensing result that relies on Gaussian measurements is non-adaptive, whereas our algorithm crucially uses {\em adaptivity} to achieve the improved sample complexity bound. We prove that adaptivity is in fact necessary in the Fourier setting: Any {\em non-adaptive} algorithm must use samples for the )-block sparse model, ruling out improvements over the vanilla sparsity assumption. Our main technical innovation for adaptivity is a new randomized energy-based importance sampling technique that may be of independent interest
An Introductory Guide to Fano's Inequality with Applications in Statistical Estimation
Information theory plays an indispensable role in the development of
algorithm-independent impossibility results, both for communication problems
and for seemingly distinct areas such as statistics and machine learning. While
numerous information-theoretic tools have been proposed for this purpose, the
oldest one remains arguably the most versatile and widespread: Fano's
inequality. In this chapter, we provide a survey of Fano's inequality and its
variants in the context of statistical estimation, adopting a versatile
framework that covers a wide range of specific problems. We present a variety
of key tools and techniques used for establishing impossibility results via
this approach, and provide representative examples covering group testing,
graphical model selection, sparse linear regression, density estimation, and
convex optimization.Comment: Chapter in upcoming book "Information-Theoretic Methods in Data
Science" (Cambridge University Press) edited by Yonina Eldar and Miguel
Rodrigues. (v2 & v3) Minor corrections and edit
An Introductory Guide to Fano's Inequality with Applications in Statistical Estimation
Information theory plays an indispensable role in the development of algorithm-independent impossibility results, both for communication problems and for seemingly distinct areas such as statistics and machine learning. While numerous information-theoretic tools have been proposed for this purpose, the oldest one remains arguably the most versatile and widespread: Fano's inequality. In this chapter, we provide a survey of Fano's inequality and its variants in the context of statistical estimation, adopting a versatile framework that covers a wide range of specic problems. We present a variety of key tools and techniques used for establishing impossibility results via this approach, and provide representative examples covering group testing, graphical model selection, sparse linear regression, density estimation, and convex optimization