2 research outputs found

    Space lower bounds for linear prediction in the streaming model

    Full text link
    We show that fundamental learning tasks, such as finding an approximate linear separator or linear regression, require memory at least \emph{quadratic} in the dimension, in a natural streaming setting. This implies that such problems cannot be solved (at least in this setting) by scalable memory-efficient streaming algorithms. Our results build on a memory lower bound for a simple linear-algebraic problem -- finding orthogonal vectors -- and utilize the estimates on the packing of the Grassmannian, the manifold of all linear subspaces of fixed dimension.Comment: Added a minor correction in referencing the prior wor

    Exponentially Improved Dimensionality Reduction for β„“1\ell_1: Subspace Embeddings and Independence Testing

    Full text link
    Despite many applications, dimensionality reduction in the β„“1\ell_1-norm is much less understood than in the Euclidean norm. We give two new oblivious dimensionality reduction techniques for the β„“1\ell_1-norm which improve exponentially over prior ones: 1. We design a distribution over random matrices S∈RrΓ—nS \in \mathbb{R}^{r \times n}, where r=2poly(d/(Ρδ))r = 2^{\textrm{poly}(d/(\varepsilon \delta))}, such that given any matrix A∈RnΓ—dA \in \mathbb{R}^{n \times d}, with probability at least 1βˆ’Ξ΄1-\delta, simultaneously for all xx, βˆ₯SAxβˆ₯1=(1Β±Ξ΅)βˆ₯Axβˆ₯1\|SAx\|_1 = (1 \pm \varepsilon)\|Ax\|_1. Note that SS is linear, does not depend on AA, and maps β„“1\ell_1 into β„“1\ell_1. Our distribution provides an exponential improvement on the previous best known map of Wang and Woodruff (SODA, 2019), which required r=22Ξ©(d)r = 2^{2^{\Omega(d)}}, even for constant Ξ΅\varepsilon and Ξ΄\delta. Our bound is optimal, up to a polynomial factor in the exponent, given a known 2d2^{\sqrt d} lower bound for constant Ξ΅\varepsilon and Ξ΄\delta. 2. We design a distribution over matrices S∈RkΓ—nS \in \mathbb{R}^{k \times n}, where k=2O(q2)(Ξ΅βˆ’1qlog⁑d)O(q)k = 2^{O(q^2)}(\varepsilon^{-1} q \log d)^{O(q)}, such that given any qq-mode tensor A∈(Rd)βŠ—qA \in (\mathbb{R}^{d})^{\otimes q}, one can estimate the entrywise β„“1\ell_1-norm βˆ₯Aβˆ₯1\|A\|_1 from S(A)S(A). Moreover, S=S1βŠ—S2βŠ—β‹―βŠ—SqS = S^1 \otimes S^2 \otimes \cdots \otimes S^q and so given vectors u1,…,uq∈Rdu_1, \ldots, u_q \in \mathbb{R}^d, one can compute S(u1βŠ—u2βŠ—β‹―βŠ—uq)S(u_1 \otimes u_2 \otimes \cdots \otimes u_q) in time 2O(q2)(Ξ΅βˆ’1qlog⁑d)O(q)2^{O(q^2)}(\varepsilon^{-1} q \log d)^{O(q)}, which is much faster than the dqd^q time required to form u1βŠ—u2βŠ—β‹―βŠ—uqu_1 \otimes u_2 \otimes \cdots \otimes u_q. Our linear map gives a streaming algorithm for independence testing using space 2O(q2)(Ξ΅βˆ’1qlog⁑d)O(q)2^{O(q^2)}(\varepsilon^{-1} q \log d)^{O(q)}, improving the previous doubly exponential (Ξ΅βˆ’1log⁑d)qO(q)(\varepsilon^{-1} \log d)^{q^{O(q)}} space bound of Braverman and Ostrovsky (STOC, 2010).Comment: To appear in COLT 2021; v2: minor fixes for camera ready versio
    corecore