Optimal Software Pipelining and Warp Specialization for Tensor Core GPUs
Stanford University
arXiv:2512.18134 [cs.PL], (19 Dec 2025)
@misc{soi2025optimalsoftwarepipeliningwarp,
title={Optimal Software Pipelining and Warp Specialization for Tensor Core GPUs},
author={Rupanshu Soi and Rohan Yadav and Fredrik Kjolstad and Alex Aiken and Maryam Mehri Dehnavi and Michael Garland and Michael Bauer},
year={2025},
eprint={2512.18134},
archivePrefix={arXiv},
primaryClass={cs.PL},
url={https://arxiv.org/abs/2512.18134}
}
GPU architectures have continued to grow in complexity, with recent incarnations introducing increasingly powerful fixed-function units for matrix multiplication and data movement to accompany highly parallel general-purpose cores. To fully leverage these machines, software must use sophisticated schedules that maximally utilize all hardware resources. Since realizing such schedules is complex, both programmers and compilers routinely employ program transformations, such as software pipelining (SWP) and warp specialization (WS), to do so in practice. However, determining how best to use SWP and WS in combination is a challenging problem that is currently handled through a mix of brittle compilation heuristics and fallible human intuition, with little insight into the space of solutions. To remedy this situation, we introduce a novel formulation of SWP and WS as a joint optimization problem that can be solved holistically by off-the-shelf constraint solvers. We reify our approach in Twill, the first system that automatically derives optimal SWP and WS schedules for a large class of iterative programs. Twill is heuristic-free, easily extensible to new GPU architectures, and guaranteed to produce optimal schedules. We show that Twill can rediscover, and thereby prove optimal, the SWP and WS schedules manually developed by experts for Flash Attention on both the NVIDIA Hopper and Blackwell GPU architectures.
December 29, 2025 by hgpu
Your response
You must be logged in to post a comment.




