ML Inference Scheduling with Predictable Latency
Inria & Sorbonne University, Paris, France
Proceedings of the Middleware for Autonomous AIoT Systems in the Computing Continuum (MAIoT ’25), 2025
@incollection{zhao2025ml,
title={ML Inference Scheduling with Predictable Latency},
author={Zhao, Haidong and Georgantas, Nikolaos},
booktitle={Proceedings of the Middleware for Autonomous AIoT Systems in the Computing Continuum},
pages={25–30},
year={2025}
}
Machine learning (ML) inference serving systems can schedule requests to improve GPU utilization and to meet service level objectives (SLOs) or deadlines. However, improving GPU utilization may compromise latency-sensitive scheduling, as concurrent tasks contend for GPU resources and thereby introduce interference. Given that interference effects introduce unpredictability in scheduling, neglecting them may compromise SLO or deadline satisfaction. Nevertheless, existing interference prediction approaches remain limited in several respects, which may restrict their usefulness for scheduling. First, they are often coarse-grained, which ignores runtime co-location dynamics and thus restricts their accuracy in interference prediction. Second, they tend to use a static prediction model, which may not effectively cope with different workload characteristics. To this end, we evaluate the potential limitations of existing interference prediction approaches and outline our ongoing work toward achieving efficient ML inference scheduling.
December 21, 2025 by hgpu
Your response
You must be logged in to post a comment.




