Skip to main content
Article thumbnail
Location of Repository

Communication Optimal Parallel Multiplication of Sparse Random Matrices ∗

By Grey Ballard, Laura Grigori, Aydın Buluç, Benjamin Lipshitz, Sivan Toledo, James Demmel and Oded Schwartz

Abstract

Parallel algorithms for sparse matrix-matrix multiplication typically spend most of their time on inter-processor communication rather than on computation, and hardware trends predict the relative cost of communication will only increase. Thus, sparse matrix multiplication algorithms must minimize communication costs in order to scale to large processor counts. In this paper, we consider multiplying sparse matrices corresponding to Erdős-Rényi random graphs on distributedmemory parallel machines. We prove a new lower bound on the expected communication cost for a wide class of algorithms. Our analysis of existing algorithms shows that, while some are optimal for a limited range of matrix density and number of processors, none is optimal in general. We obtain two new parallel algorithms and prove that they match the expected communication cost lower bound, and hence they are optimal. We acknowledge funding from Microsoft (Award #024263

Year: 2013
OAI identifier: oai:CiteSeerX.psu:10.1.1.352.4321
Provided by: CiteSeerX
Download PDF:
Sorry, we are unable to provide the full text but you may find it at the following location(s):
  • http://citeseerx.ist.psu.edu/v... (external link)
  • http://gauss.cs.ucsb.edu/~aydi... (external link)
  • Suggested articles


    To submit an update or takedown request for this paper, please submit an Update/Correction/Removal Request.