Autotuning GPU Kernels via Static and Predictive Analysis
Computer and Information Science, University of Oregon, Eugene, OR, USA
arXiv:1701.08547 [cs.DC], (30 Jan 2017)
@article{lim2017autotuning,
title={Autotuning GPU Kernels via Static and Predictive Analysis},
author={Lim, Robert V. and Norris, Boyana and Malony, Allen D.},
year={2017},
month={jan},
archivePrefix={"arXiv"},
primaryClass={cs.DC}
}
Optimizing the performance of GPU kernels is challenging for both human programmers and code generators. For example, CUDA programmers must set thread and block parameters for a kernel, but might not have the intuition to make a good choice. Similarly, compilers can generate working code, but may miss tuning opportunities by not targeting GPU models or performing code transformations. Although empirical autotuning addresses some of these challenges, it requires extensive experimentation and search for optimal code variants. This research presents an approach for tuning CUDA kernels based on static analysis that considers fine-grained code structure and the specific GPU architecture features. Notably, our approach does not require any program runs in order to discover near-optimal parameter settings. We demonstrate the applicability of our approach in enabling code autotuners such as Orio to produce competitive code variants comparable with empirical-based methods, without the high cost of experiments.
February 2, 2017 by hgpu