Beyond Code Pairs: Dialogue-Based Data Generation for LLM Code Translation
Argonne National Laboratory, Lemont, USA
arXiv:2512.03086 [cs.PL], (29 Nov 2025)
@misc{chen2025codepairsdialoguebaseddata,
title={Beyond Code Pairs: Dialogue-Based Data Generation for LLM Code Translation},
author={Le Chen and Nuo Xu and Winson Chen and Bin Lei and Pei-Hung Lin and Dunzhi Zhou and Rajeev Thakur and Caiwen Ding and Ali Jannesari and Chunhua Liao},
year={2025},
eprint={2512.03086},
archivePrefix={arXiv},
primaryClass={cs.PL},
url={https://arxiv.org/abs/2512.03086}
}
Large language models (LLMs) have shown remarkable capabilities in code translation, yet their performance deteriorates in low-resource programming domains such as Fortran and emerging frameworks like CUDA, where high-quality parallel data are scarce. We present an automated dataset generation pipeline featuring a dual-LLM Questioner-Solver design that incorporates external knowledge from compilers and runtime feedback. Beyond traditional source-target code pair datasets, our approach additionally generates (1) verified translations with unit tests for assessing functional consistency, and (2) multi-turn dialogues that capture the reasoning process behind translation refinement. Applied to Fortran -> C++ and C++ -> CUDA, the pipeline yields 3.64k and 3.93k dialogues, respectively. Fine-tuning on this data yields dramatic improvements in functional correctness, boosting unit test success rates by over 56% on the challenging C++-to-CUDA task. We show this data enables a 7B open-weight model to significantly outperform larger proprietary systems on key metrics like compilation success.
December 21, 2025 by hgpu
Your response
You must be logged in to post a comment.




