KernelBand: Boosting LLM-based Kernel Optimization with a Hierarchical and Hardware-aware Multi-armed Bandit
Key Lab of HCST (PKU), MOE; SCS, Peking University, Beijing, China
arXiv:2511.18868 [cs.LG]
@misc{ran2025kernelbandboostingllmbasedkernel,
title={KernelBand: Boosting LLM-based Kernel Optimization with a Hierarchical and Hardware-aware Multi-armed Bandit},
author={Dezhi Ran and Shuxiao Xie and Mingfang Ji and Ziyue Hua and Mengzhou Wu and Yuan Cao and Yuzhe Guo and Yu Hao and Linyi Li and Yitao Hu and Tao Xie},
year={2025},
eprint={2511.18868},
archivePrefix={arXiv},
primaryClass={cs.LG},
url={https://arxiv.org/abs/2511.18868}
}
High quality kernels are critical for reducing training and inference costs of Large Language Models (LLMs), yet they traditionally require significant expertise in hardware architecture and software optimization. While recent advances in LLM-based code generation show promise for complex optimization, existing methods struggle with the vast optimization space due to insufficient hardware domain knowledge, failing to effectively balance exploration and exploitation. We present KernelBand, a novel framework that formulates kernel optimization as a hierarchical multi-armed bandit problem, enabling LLM agents to strategically navigate the optimization space by treating kernel selection and optimization strategy application as sequential decision-making processes. Our approach leverages hardware profiling information to identify promising optimization strategies and employs runtime behavior clustering to reduce exploration overhead across kernel candidates. Extensive experiments on TritonBench demonstrate that KernelBand significantly outperforms state-of-the-art methods, achieving superior performance with fewer tokens while exhibiting consistent improvement without saturation as computational resources increase.
November 30, 2025 by hgpu
Your response
You must be logged in to post a comment.




