21 research outputs found

    Computational Complexity of Certifying Restricted Isometry Property

    Get PDF
    Given a matrix AA with nn rows, a number k<nk<n, and 0<δ<10<\delta < 1, AA is (k,δ)(k,\delta)-RIP (Restricted Isometry Property) if, for any vector xRnx \in \mathbb{R}^n, with at most kk non-zero co-ordinates, (1δ)x2Ax2(1+δ)x2(1-\delta) \|x\|_2 \leq \|A x\|_2 \leq (1+\delta)\|x\|_2 In many applications, such as compressed sensing and sparse recovery, it is desirable to construct RIP matrices with a large kk and a small δ\delta. Given the efficacy of random constructions in generating useful RIP matrices, the problem of certifying the RIP parameters of a matrix has become important. In this paper, we prove that it is hard to approximate the RIP parameters of a matrix assuming the Small-Set-Expansion-Hypothesis. Specifically, we prove that for any arbitrarily large constant C>0C>0 and any arbitrarily small constant 0<δ<10<\delta<1, there exists some kk such that given a matrix MM, it is SSE-Hard to distinguish the following two cases: - (Highly RIP) MM is (k,δ)(k,\delta)-RIP. - (Far away from RIP) MM is not (k/C,1δ)(k/C, 1-\delta)-RIP. Most of the previous results on the topic of hardness of RIP certification only hold for certification when δ=o(1)\delta=o(1). In practice, it is of interest to understand the complexity of certifying a matrix with δ\delta being close to 21\sqrt{2}-1, as it suffices for many real applications to have matrices with δ=21\delta = \sqrt{2}-1. Our hardness result holds for any constant δ\delta. Specifically, our result proves that even if δ\delta is indeed very small, i.e. the matrix is in fact \emph{strongly RIP}, certifying that the matrix exhibits \emph{weak RIP} itself is SSE-Hard. In order to prove the hardness result, we prove a variant of the Cheeger's Inequality for sparse vectors

    Computational barriers in minimax submatrix detection

    Get PDF
    This paper studies the minimax detection of a small submatrix of elevated mean in a large matrix contaminated by additive Gaussian noise. To investigate the tradeoff between statistical performance and computational cost from a complexity-theoretic perspective, we consider a sequence of discretized models which are asymptotically equivalent to the Gaussian model. Under the hypothesis that the planted clique detection problem cannot be solved in randomized polynomial time when the clique size is of smaller order than the square root of the graph size, the following phase transition phenomenon is established: when the size of the large matrix pp\to\infty, if the submatrix size k=Θ(pα)k=\Theta(p^{\alpha}) for any α(0,2/3)\alpha\in(0,{2}/{3}), computational complexity constraints can incur a severe penalty on the statistical performance in the sense that any randomized polynomial-time test is minimax suboptimal by a polynomial factor in pp; if k=Θ(pα)k=\Theta(p^{\alpha}) for any α(2/3,1)\alpha\in({2}/{3},1), minimax optimal detection can be attained within constant factors in linear time. Using Schatten norm loss as a representative example, we show that the hardness of attaining the minimax estimation rate can crucially depend on the loss function. Implications on the hardness of support recovery are also obtained.Comment: Published at http://dx.doi.org/10.1214/14-AOS1300 in the Annals of Statistics (http://www.imstat.org/aos/) by the Institute of Mathematical Statistics (http://www.imstat.org

    Finding Planted Cliques in Sublinear Time

    Full text link
    We study the planted clique problem in which a clique of size kk is planted in an Erd\H{o}s-R\'enyi graph of size nn and one wants to recover this planted clique. For k=Ω(n)k=\Omega(\sqrt{n}), polynomial time algorithms can find the planted clique. The fastest such algorithms run in time linear O(n2)O(n^2) (or nearly linear) in the size of the input [FR10,DGGP14,DM15a]. In this work, we initiate the development of sublinear time algorithms that find the planted clique when k=ω(nloglogn)k=\omega(\sqrt{n \log \log n}). Our algorithms can recover the clique in time O~(n+(nk)3)=O~(n32)\widetilde{O}\left(n+(\frac{n}{k})^{3}\right)=\widetilde{O}\left(n^{\frac{3}{2}}\right) when k=Ω(nlogn)k=\Omega(\sqrt{n\log n}), and in time O~(n2/exp(k224n))\widetilde{O}\left(n^2/\exp{\left(\frac{k^2}{24n}\right)}\right) for ω(nloglogn)=k=o(nlogn)\omega(\sqrt{n\log \log n})=k=o(\sqrt{n\log{n}}). An Ω(n){\Omega}(n) running time lower bound for the planted clique recovery problem follows easily from the results of [RS19] and therefore our recovery algorithms are optimal whenever k=Ω(n23)k = \Omega(n^{\frac{2}{3}}). As the lower bound of [RS19] builds on purely information theoretic arguments, it cannot provide a detection lower bound stronger than Ω~(n2k2)\widetilde{\Omega}(\frac{n^2}{k^2}). Since our algorithms for k=Ω(nlogn)k = \Omega(\sqrt{n \log n}) run in time O~(n3k3+n)\widetilde{O}\left(\frac{n^3}{k^3} + n\right), we show stronger lower bounds based on computational hardness assumptions. With a slightly different notion of the planted clique problem we show that the Planted Clique Conjecture implies the following. A natural family of non-adaptive algorithms---which includes our algorithms for clique detection---cannot reliably solve the planted clique detection problem in time O(n3δk3)O\left( \frac{n^{3-\delta}}{k^3}\right) for any constant δ>0\delta>0. Thus we provide evidence that if detecting small cliques is hard, it is also likely that detecting large cliques is not \textit{too} easy
    corecore