TritonForge: Profiling-Guided Framework for Automated Triton Kernel Optimization
University of California, Riverside
arXiv:2512.09196 [cs.SE], (9 Dec 2025)
@misc{li2025tritonforgeprofilingguidedframeworkautomated,
title={TritonForge: Profiling-Guided Framework for Automated Triton Kernel Optimization},
author={Haonan Li and Keyu Man and Partha Kanuparthy and Hanning Chen and Wei Sun and Sreen Tallam and Chenguang Zhu and Kevin Zhu and Zhiyun Qian},
year={2025},
eprint={2512.09196},
archivePrefix={arXiv},
primaryClass={cs.SE},
url={https://arxiv.org/abs/2512.09196}
}
High-performance GPU kernel optimization remains a critical yet labor-intensive task in modern machine learning workloads. Although Triton, a domain-specific language for GPU programming, enables developers to write efficient kernels with concise code, achieving expert-level performance still requires deep understanding of GPU architectures and low-level performance trade-offs. We present TritonForge, a profiling-guided framework for automated Triton kernel optimization. TritonForge integrates kernel analysis, runtime profiling, and iterative code transformation to streamline the optimization process. By incorporating data-driven feedback from profiling results, the system identifies performance bottlenecks, proposes targeted code modifications, and evaluates their impact automatically. While our prototype leverages large language models (LLMs) to assist in code reasoning and transformation, the framework remains modular and model-agnostic. Across diverse kernel types and GPU architectures, TritonForge achieves up to 5x performance improvement over baseline implementations and on average 1.76x of the cases are successful, providing a foundation for future research in automated GPU performance optimization.
December 15, 2025 by hgpu
Your response
You must be logged in to post a comment.





