9680

A model of dynamic compilation for heterogeneous compute platforms

Andrew Kerr
Georgia Institute of Technology
Georgia Institute of Technology, 2013

@article{kerr2013model,

   title={A model of dynamic compilation for heterogeneous compute platforms},

   author={Kerr, Andrew},

   year={2013},

   publisher={Georgia Institute of Technology}

}

Download Download (PDF)   View View   Source Source   

2027

views

Trends in computer engineering place renewed emphasis on increasing parallelism and heterogeneity. The rise of parallelism adds an additional dimension to the challenge of portability, as different processors support different notions of parallelism, whether vector parallelism executing in a few threads on multicore CPUs or large-scale thread hierarchies on GPUs. Thus, software experiences obstacles to portability and efficient execution beyond differences in instruction sets; rather, the underlying execution models of radically different architectures may not be compatible. Dynamic compilation applied to data-parallel heterogeneous architectures presents an abstraction layer decoupling program representations from optimized binaries, thus enabling portability without encumbering performance. This dissertation proposes several techniques that extend dynamic compilation to data-parallel execution models. These contributions include: – characterization of data-parallel workloads; – machine-independent application metrics; – framework for performance modeling and prediction; – execution model translation for vector processors; – region-based compilation and scheduling. We evaluate these claims via the development of a novel dynamic compilation framework, GPU Ocelot, with which we execute real-world workloads from GPU computing. This enables the execution of GPU computing workloads to run efficiently on multicore CPUs, GPUs, and a functional simulator. We show data-parallel workloads exhibit performance scaling, take advantage of vector instruction set extensions, and effectively exploit data locality via scheduling which attempts to maximize control locality.
No votes yet.
Please wait...

* * *

* * *

HGPU group © 2010-2024 hgpu.org

All rights belong to the respective authors

Contact us: