A GEMM interface and implementation on NVIDIA GPUs for multiple small matrices
Tech-X Corporation, Boulder
@article{tgemm,
author={Chetan Jhurani and Paul Mullowney},
title={A GEMM interface and implementation on NVIDIA GPUs for multiple small matrices},
journal={Submitted},
url={\url{www.ices.utexas.edu/$\char126$chetan/preprints/2013-CJ-PM-GEMM.pdf}},
year={2013}
}
We present an interface and an implementation of the General Matrix Multiply (GEMM) routine for multiple small matrices processed simultaneously on NVIDIA graphics processing units (GPUs). We focus on matrix sizes under 16. The implementation can be easily extended to larger sizes. For single precision matrices, our implementation is 30% to 600% faster than the batched cuBLAS implementation distributed in the CUDA Toolkit 5.0 on NVIDIA Tesla K20c. For example, we obtain 104 GFlop/s and 216 GFlop/s when multiplying 100,000 independent matrix pairs of size 10 and 16, respectively. Similar improvement in performance is obtained for other sizes, in single and double precision for real and complex types, and when the number of matrices is smaller. Apart from our implementation, our different function interface also plays an important role in the improved performance. Applications of this software include Finite Element computation on GPUs.
April 9, 2013 by chetan.jhurani