2 research outputs found
A hardware acceleration technique for gradient descent and conjugate gradient
Gradient descent, conjugate gradient, and other iterative algorithms are a
powerful class of algorithms; however, they can take a long time for conver-
gence. Baseline accelerator designs feature insu cient coverage of operations
and do not work well on the problems we target. In this thesis we present
a novel hardware architecture for accelerating gradient descent and other
similar algorithms. To support this architecture, we also present a sparse
matrix-vector storage format, and software support for utilizing the format,
so that it can be e ciently mapped onto hardware which is also well suited for
dense operations. We show that the accelerator design outperforms similar
designs which target only the most dominant operation of a given algorithm,
providing substantial energy and performance bene ts. We further show that
the accelerator can be reasonably implemented on a general purpose CPU
with small area overhead