30440

Targeted Testing of Compiler Optimizations via Grammar-Level Composition Styles

Zitong Zhou, Ben Limpanukorn, Hong Jin Kang, Jiyuan Wang, Yaoxuan Wu, Akos Kiss, Renata Hodovan, Miryung Kim
University of California, Los Angeles (UCLA), USA
arXiv:2512.04344 [cs.SE], (4 Dec 2025)

@misc{zhou2025targetedtestingcompileroptimizations,

   title={Targeted Testing of Compiler Optimizations via Grammar-Level Composition Styles},

   author={Zitong Zhou and Ben Limpanukorn and Hong Jin Kang and Jiyuan Wang and Yaoxuan Wu and Akos Kiss and Renata Hodovan and Miryung Kim},

   year={2025},

   eprint={2512.04344},

   archivePrefix={arXiv},

   primaryClass={cs.SE},

   url={https://arxiv.org/abs/2512.04344}

}

Download Download (PDF)   View View   Source Source   

168

views

Ensuring the correctness of compiler optimizations is critical, but existing fuzzers struggle to test optimizations effectively. First, most fuzzers use optimization pipelines (heuristics-based, fixed sequences of passes) as their harness. The phase-ordering problem can enable or preempt transformations, so pipelines inevitably miss optimization interactions; moreover, many optimizations are not scheduled, even at aggressive levels. Second, optimizations typically fire only when inputs satisfy specific structural relationships, which existing generators and mutations struggle to produce. We propose targeted fuzzing of individual optimizations to complement pipeline-based testing. Our key idea is to exploit composition styles – structural relations over program constructs (adjacency, nesting, repetition, ordering) – that optimizations look for. We build a general-purpose, grammar-based mutational fuzzer, TargetFuzz, that (i) mines composition styles from an optimization-relevant corpus, then (ii) rebuilds them inside different contexts offered by a larger, generic corpus via synthesized mutations to test variations of optimization logic. TargetFuzz is adaptable to a new programming language by lightweight, grammar-based, construct annotations – and it automatically synthesizes mutators and crossovers to rebuild composition styles. No need for hand-coded generators or language-specific mutators, which is particularly useful for modular frameworks such as MLIR, whose dialect-based, rapidly evolving ecosystem makes optimizations difficult to fuzz. Our evaluation on LLVM and MLIR shows that TargetFuzz improves coverage by 8% and 11% and triggers optimizations 2.8 and 2.6, compared to baseline fuzzers under the targeted fuzzing mode. We show that targeted fuzzing is complementary: it effectively tests all 37 sampled LLVM optimizations, while pipeline-fuzzing missed 12.
No votes yet.
Please wait...

You must be logged in to post a comment.

* * *

* * *

HGPU group © 2010-2025 hgpu.org

All rights belong to the respective authors

Contact us: