26261

A ML-based resource utilization OpenCL GPU-kernel fusion model

Usman Ahmed, Jerry Chun-Wei Lin, Gautam Srivastava
Department of Computer Science, Electrical Engineering and Mathematical Sciences, Western Norway University of Applied Sciences, Bergen, Norway
Sustainable Computing: Informatics and Systems, Volume 35, 100683, 2022

@article{ahmed2022ml,

   title={A ML-based resource utilization OpenCL GPU-kernel fusion model},

   author={Ahmed, Usman and Lin, Jerry Chun-Wei and Srivastava, Gautam},

   journal={Sustainable Computing: Informatics and Systems},

   pages={100683},

   year={2022},

   publisher={Elsevier}

}

Download Download (PDF)   View View   Source Source   

1061

views

Massive data parallelism can be achieved by using general-purpose graphics processing units (GPGPU) with the help of the OpenCL framework. When smaller data with higher GPU memory is executed, it results in a low resource utilization ratio and energy inefficiencies. Up until now, there is no existing model to share GPU for further execution. In addition, if the kernel pair requires the same computation resource, then kernel merging also results in a significant increase in execution time. Therefore, optimal device selection, as well as kernel merging, can significantly speed up the execution performance for a batch of jobs. This paper proposes a kernel merging method that leads to high GPU occupancy. As a result, it reduces execution time and increases GPU utilization. Additionally, a machine learning (ML)-based GPU sharing mechanism is presented to select pairs of kernels in OpenCL frameworks. The model first selects suitable architecture for the jobs and then merges GPU kernels for better resource utilization. From all the GPU candidates, the optimal pair of the kernel concerning data size is selected. The experimental results show that the developed model can achieve 0.91 F1-measure for device selection and 0.98 for the scheduling scheme of kernel merging.
No votes yet.
Please wait...

* * *

* * *

HGPU group © 2010-2024 hgpu.org

All rights belong to the respective authors

Contact us: