Comparison of GPU Architectures for Asynchronous Communication with Finite-Differencing Applications

D.P. Playne, K.A. Hawick
Institute of Information and Mathematical Sciences, Massey University – Albany, North Shore 102-904, Auckland, New Zealand.
Concurrency and Computation: Practice and Experience, 2011


   title={Comparison of GPU architectures for asynchronous communication with finite-differencing applications},

   author={Playne, DP and Hawick, KA},


   publisher={Wiley Online Library},

   journal={Concurrency and Computation: Practice and Experience}


Download Download (PDF)   View View   Source Source   



Graphical Processing Units (GPUs) are good data-parallel performance accelerators for solving regular mesh partial differential equations (PDEs) whereby low-latency communications and high compute to communications ratios can yield very high levels of computational efficiency. Finite-difference time-domain methods still play an important role for many PDE applications. Iterative multi-grid and multilevel algorithms can converge faster than ordinary finite difference methods but can be much more difficult to parallelise with GPU memory constraints. We report on some practical algorithmic and data layout approaches and on performance data on a range of GPUs with CUDA. We focus on the use of multiple GPU devices with a single CPU host and the asynchronous CPU/GPU communications issues involved. We obtain more than two orders of magnitude of speedup over a comparable CPU core.
No votes yet.
Please wait...

* * *

* * *

HGPU group © 2010-2021 hgpu.org

All rights belong to the respective authors

Contact us: