30340

AMD MI300X GPU Performance Analysis

Chandrish Ambati, Trung Diep
Celestial AI
arXiv:2510.27583 [cs.PF], (31 Oct 2025)

@misc{ambati2025amdmi300xgpuperformance,

   title={AMD MI300X GPU Performance Analysis},

   author={Chandrish Ambati and Trung Diep},

   year={2025},

   eprint={2510.27583},

   archivePrefix={arXiv},

   primaryClass={cs.PF},

   url={https://arxiv.org/abs/2510.27583}

}

Download Download (PDF)   View View   Source Source   

236

views

The rapid growth of large language models (LLMs) has driven the need for high-performance, scalable GPU hardware capable of efficiently serving models with hundreds of billions of parameters. While NVIDIA GPUs have traditionally dominated LLM deployments due to their mature CUDA software stack and state-of the-art accelerators, AMD’s latest MI300X GPUs offer a compelling alternative, featuring high HBM capacity, matrix cores, and their proprietary interconnect. In this paper, we present a comprehensive evaluation of the AMD MI300X GPUs across key performance domains critical to LLM inference including compute throughput, memory bandwidth, and interconnect communication.
No votes yet.
Please wait...

You must be logged in to post a comment.

* * *

* * *

HGPU group © 2010-2025 hgpu.org

All rights belong to the respective authors

Contact us: