SQ Lower Bounds for Learning Mixtures of Linear Classifiers

Abstract

We study the problem of learning mixtures of linear classifiers under Gaussian covariates. Given sample access to a mixture of rr distributions on Rn\mathbb{R}^n of the form (x,yβ„“)(\mathbf{x},y_{\ell}), β„“βˆˆ[r]\ell\in [r], where x∼N(0,In)\mathbf{x}\sim\mathcal{N}(0,\mathbf{I}_n) and yβ„“=sign(⟨vβ„“,x⟩)y_\ell=\mathrm{sign}(\langle\mathbf{v}_\ell,\mathbf{x}\rangle) for an unknown unit vector vβ„“\mathbf{v}_\ell, the goal is to learn the underlying distribution in total variation distance. Our main result is a Statistical Query (SQ) lower bound suggesting that known algorithms for this problem are essentially best possible, even for the special case of uniform mixtures. In particular, we show that the complexity of any SQ algorithm for the problem is npoly(1/Ξ”)log⁑(r)n^{\mathrm{poly}(1/\Delta) \log(r)}, where Ξ”\Delta is a lower bound on the pairwise β„“2\ell_2-separation between the vβ„“\mathbf{v}_\ell's. The key technical ingredient underlying our result is a new construction of spherical designs that may be of independent interest.Comment: To appear in NeurIPS 202

    Similar works

    Full text

    thumbnail-image

    Available Versions