SQ Lower Bounds for Learning Bounded Covariance GMMs

Abstract

We study the complexity of learning mixtures of separated Gaussians with common unknown bounded covariance matrix. Specifically, we focus on learning Gaussian mixture models (GMMs) on Rd\mathbb{R}^d of the form P=i=1kwiN(μi,Σi)P= \sum_{i=1}^k w_i \mathcal{N}(\boldsymbol \mu_i,\mathbf \Sigma_i), where Σi=ΣI\mathbf \Sigma_i = \mathbf \Sigma \preceq \mathbf I and minijμiμj2kϵ\min_{i \neq j} \| \boldsymbol \mu_i - \boldsymbol \mu_j\|_2 \geq k^\epsilon for some ϵ>0\epsilon>0. Known learning algorithms for this family of GMMs have complexity (dk)O(1/ϵ)(dk)^{O(1/\epsilon)}. In this work, we prove that any Statistical Query (SQ) algorithm for this problem requires complexity at least dΩ(1/ϵ)d^{\Omega(1/\epsilon)}. In the special case where the separation is on the order of k1/2k^{1/2}, we additionally obtain fine-grained SQ lower bounds with the correct exponent. Our SQ lower bounds imply similar lower bounds for low-degree polynomial tests. Conceptually, our results provide evidence that known algorithms for this problem are nearly best possible

    Similar works

    Full text

    thumbnail-image

    Available Versions