Optimal Workload Placement on Multi-Instance GPUs
IBM Research, United States
arXiv:2409.06646 [cs.DC], (10 Sep 2024)
@misc{turkkan2024optimalworkloadplacementmultiinstance,
title={Optimal Workload Placement on Multi-Instance GPUs},
author={Bekir Turkkan and Pavankumar Murali and Pavithra Harsha and Rohan Arora and Gerard Vanloo and Chandra Narayanaswami},
year={2024},
eprint={2409.06646},
archivePrefix={arXiv},
primaryClass={cs.DC},
url={https://arxiv.org/abs/2409.06646}
}
There is an urgent and pressing need to optimize usage of Graphical Processing Units (GPUs), which have arguably become one of the most expensive and sought after IT resources. To help with this goal, several of the current generation of GPUs support a partitioning feature, called Multi-Instance GPU (MIG) to allow multiple workloads to share a GPU, albeit with some constraints. In this paper we investigate how to optimize the placement of Large Language Model (LLM)-based AI Inferencing workloads on GPUs. We first identify and present several use cases that are encountered in practice that require workloads to be efficiently placed or migrated to other GPUs to make room for incoming workloads. The overarching goal is to use as few GPUs as possible and to further minimize memory and compute wastage on GPUs that are utilized. We have developed two approaches to address this problem: an optimization method and a heuristic method. We benchmark these with two workload scheduling heuristics for multiple use cases. Our results show up to 2.85x improvement in the number of GPUs used and up to 70% reduction in GPU wastage over baseline heuristics. We plan to enable the SRE community to leverage our proposed method in production environments.
September 15, 2024 by hgpu