30188

High Performance Matrix Multiplication

Ethan Davis
arXiv:2509.04594 [cs.PF], (4 Sep 2025)

@misc{davis2025high,

   title={High Performance Matrix Multiplication},

   author={Ethan Davis},

   year={2025},

   eprint={2509.04594},

   archivePrefix={arXiv},

   primaryClass={cs.PF}

}

Matrix multiplication is the foundation from much of the success from high performance technologies like deep learning, scientific simulations, and video graphics. High level programming languages like Python and R rely on highly optimized low level libraries for performing core linear algebra operations like matrix multiplication from Basic Linear Algebra Subprograms (BLAS). This paper compares the performance of five different matrix multiplication algorithms using CuBLAS, CUDA, BLAS, OpenMP, and C++ Threads. We find statistical significance with a p-value below 5e-12 to support the hypothesis that for square matrices where is at least 10,000 then the in order performance as measured in floating point operations per second (FLOPS) for these matrix multiplication algorithms is CuBLAS, CUDA, BLAS, OpenMP, and C++ Threads.
No votes yet.
Please wait...

You must be logged in to post a comment.

* * *

* * *

HGPU group © 2010-2025 hgpu.org

All rights belong to the respective authors

Contact us: