3 research outputs found

    Total Least Squares Regression in Input Sparsity Time

    Full text link
    In the total least squares problem, one is given an mΓ—nm \times n matrix AA, and an mΓ—dm \times d matrix BB, and one seeks to "correct" both AA and BB, obtaining matrices A^\hat{A} and B^\hat{B}, so that there exists an XX satisfying the equation A^X=B^\hat{A}X = \hat{B}. Typically the problem is overconstrained, meaning that m≫max⁑(n,d)m \gg \max(n,d). The cost of the solution A^,B^\hat{A}, \hat{B} is given by βˆ₯Aβˆ’A^βˆ₯F2+βˆ₯Bβˆ’B^βˆ₯F2\|A-\hat{A}\|_F^2 + \|B - \hat{B}\|_F^2. We give an algorithm for finding a solution XX to the linear system A^X=B^\hat{A}X=\hat{B} for which the cost βˆ₯Aβˆ’A^βˆ₯F2+βˆ₯Bβˆ’B^βˆ₯F2\|A-\hat{A}\|_F^2 + \|B-\hat{B}\|_F^2 is at most a multiplicative (1+Ο΅)(1+\epsilon) factor times the optimal cost, up to an additive error Ξ·\eta that may be an arbitrarily small function of nn. Importantly, our running time is O~(nnz(A)+nnz(B))+poly(n/Ο΅)β‹…d\tilde{O}( \mathrm{nnz}(A) + \mathrm{nnz}(B) ) + \mathrm{poly}(n/\epsilon) \cdot d, where for a matrix CC, nnz(C)\mathrm{nnz}(C) denotes its number of non-zero entries. Importantly, our running time does not directly depend on the large parameter mm. As total least squares regression is known to be solvable via low rank approximation, a natural approach is to invoke fast algorithms for approximate low rank approximation, obtaining matrices A^\hat{A} and B^\hat{B} from this low rank approximation, and then solving for XX so that A^X=B^\hat{A}X = \hat{B}. However, existing algorithms do not apply since in total least squares the rank of the low rank approximation needs to be nn, and so the running time of known methods would be at least mn2mn^2. In contrast, we are able to achieve a much faster running time for finding XX by never explicitly forming the equation A^X=B^\hat{A} X = \hat{B}, but instead solving for an XX which is a solution to an implicit such equation. Finally, we generalize our algorithm to the total least squares problem with regularization

    A near-optimal algorithm for approximating the John Ellipsoid

    Full text link
    We develop a simple and efficient algorithm for approximating the John Ellipsoid of a symmetric polytope. Our algorithm is near optimal in the sense that our time complexity matches the current best verification algorithm. We also provide the MATLAB code for further research.Comment: COLT 201

    Sketching Transformed Matrices with Applications to Natural Language Processing

    Full text link
    Suppose we are given a large matrix A=(ai,j)A=(a_{i,j}) that cannot be stored in memory but is in a disk or is presented in a data stream. However, we need to compute a matrix decomposition of the entry-wisely transformed matrix, f(A):=(f(ai,j))f(A):=(f(a_{i,j})) for some function ff. Is it possible to do it in a space efficient way? Many machine learning applications indeed need to deal with such large transformed matrices, for example word embedding method in NLP needs to work with the pointwise mutual information (PMI) matrix, while the entrywise transformation makes it difficult to apply known linear algebraic tools. Existing approaches for this problem either need to store the whole matrix and perform the entry-wise transformation afterwards, which is space consuming or infeasible, or need to redesign the learning method, which is application specific and requires substantial remodeling. In this paper, we first propose a space-efficient sketching algorithm for computing the product of a given small matrix with the transformed matrix. It works for a general family of transformations with provable small error bounds and thus can be used as a primitive in downstream learning tasks. We then apply this primitive to a concrete application: low-rank approximation. We show that our approach obtains small error and is efficient in both space and time. We complement our theoretical results with experiments on synthetic and real data.Comment: AISTATS 202
    corecore