A Compute Unified System Architecture for Graphics Clusters Incorporating Data Locality
Visualisierungsinstitut der Universitat Stuttgart, Stuttgart, Germany
IEEE Transactions on Visualization and Computer Graphics, 2008
@article{muller2008compute,
title={A compute unified system architecture for graphics clusters incorporating data locality},
author={M{\"u}ller, C. and Frey, S. and Strengert, M. and Dachsbacher, C. and Ertl, T.},
journal={IEEE Transactions on Visualization and Computer Graphics},
pages={605–617},
year={2008},
publisher={Published by the IEEE Computer Society}
}
We present a development environment for distributed GPU computing targeted for multi-GPU systems, as well as graphics clusters. Our system is based on CUDA and logically extends its parallel programming model for graphics processors to higher levels of parallelism, namely, the PCI bus and network interconnects. While the extended API mimics the full function set of current graphics hardware-including the concept of global memory-on all distribution layers, the underlying communication mechanisms are handled transparently for the application developer. To allow for high scalability, in particular for network-interconnected environments, we introduce an automatic GPU-accelerated scheduling mechanism that is aware of data locality. This way, the overall amount of transmitted data can be heavily reduced, which leads to better GPU utilization and faster execution. We evaluate the performance and scalability of our system for bus and especially network-level parallelism on typical multi-GPU systems and graphics clusters.
May 30, 2011 by hgpu