Data-Aware Task Scheduling on Multi-accelerator Based Platforms

Cedric Augonnet, Jerome Clet-Ortega, Samuel Thibault, Raymond Namyst
LaBRI, University of Bordeaux, Bordeaux, France
IEEE 16th International Conference on Parallel and Distributed Systems (ICPADS), 2010


   title={Data-Aware Task Scheduling on Multi-Accelerator based Platforms},

   author={Augonnet, C. and Clet-Ortega, J. and Thibault, S. and Namyst, R.},

   booktitle={2010 IEEE 16th International Conference on Parallel and Distributed Systems},





Download Download (PDF)   View View   Source Source   Source codes Source codes




To fully tap into the potential of heterogeneous machines composed of multicore processors and multiple accelerators, simple offloading approaches in which the main trunk of the application runs on regular cores while only specific parts are offloaded on accelerators are not sufficient. The real challenge is to build systems where the application would permanently spread across the entire machine, that is, where parallel tasks would be dynamically scheduled over the full set of available processing units. To face this challenge, we previously proposed StarPU, a runtime system capable of scheduling tasks over multicore machines equipped with GPU accelerators. StarPU uses a software virtual shared memory (VSM) that provides a highlevel programming interface and automates data transfers between processing units so as to enable a dynamic scheduling of tasks. We now present how we have extended StarPU to minimize the cost of transfers between processing units in order to efficiently cope with multi-GPU hardware configurations. To this end, our runtime system implements data prefetching based on asynchronous data transfers, and uses data transfer cost prediction to influence the decisions taken by the task scheduler. We demonstrate the relevance of our approach by benchmarking two parallel numerical algorithms using our runtime system. We obtain significant speedups and high efficiency over multicore machines equipped with multiple accelerators. We also evaluate the behaviour of these applications over clusters featuring multiple GPUs per node, showing how our runtime system can combine with MPI.
No votes yet.
Please wait...

* * *

* * *

HGPU group © 2010-2021 hgpu.org

All rights belong to the respective authors

Contact us: