High Performance Matrix Multiplication
arXiv:2509.04594 [cs.PF], (4 Sep 2025)
@misc{davis2025high,
title={High Performance Matrix Multiplication},
author={Ethan Davis},
year={2025},
eprint={2509.04594},
archivePrefix={arXiv},
primaryClass={cs.PF}
}
Matrix multiplication is the foundation from much of the success from high performance technologies like deep learning, scientific simulations, and video graphics. High level programming languages like Python and R rely on highly optimized low level libraries for performing core linear algebra operations like matrix multiplication from Basic Linear Algebra Subprograms (BLAS). This paper compares the performance of five different matrix multiplication algorithms using CuBLAS, CUDA, BLAS, OpenMP, and C++ Threads. We find statistical significance with a p-value below 5e-12 to support the hypothesis that for square matrices where is at least 10,000 then the in order performance as measured in floating point operations per second (FLOPS) for these matrix multiplication algorithms is CuBLAS, CUDA, BLAS, OpenMP, and C++ Threads.
September 14, 2025 by hgpu
Your response
You must be logged in to post a comment.