slides

Graph Expansion and Communication Costs of Fast Matrix Multiplication

Abstract

The communication cost of algorithms (also known as I/O-complexity) is shown to be closely related to the expansion properties of the corresponding computation graphs. We demonstrate this on Strassen's and other fast matrix multiplication algorithms, and obtain first lower bounds on their communication costs. In the sequential case, where the processor has a fast memory of size MM, too small to store three nn-by-nn matrices, the lower bound on the number of words moved between fast and slow memory is, for many of the matrix multiplication algorithms, Ξ©((nM)Ο‰0β‹…M)\Omega((\frac{n}{\sqrt M})^{\omega_0}\cdot M), where Ο‰0\omega_0 is the exponent in the arithmetic count (e.g., Ο‰0=lg⁑7\omega_0 = \lg 7 for Strassen, and Ο‰0=3\omega_0 = 3 for conventional matrix multiplication). With pp parallel processors, each with fast memory of size MM, the lower bound is pp times smaller. These bounds are attainable both for sequential and for parallel algorithms and hence optimal. These bounds can also be attained by many fast algorithms in linear algebra (e.g., algorithms for LU, QR, and solving the Sylvester equation)

    Similar works

    Full text

    thumbnail-image

    Available Versions

    Last time updated on 01/04/2019