Numerical Simulation of the Complex Ginzburg-Landau Equation on GPUs with CUDA
Computer Science, Institute for Information and Mathematical Sciences, Massey University, North Shore 102-904, Auckland, New Zealand
Technical Report CSTN-070, Massey University, 2010
@conference{hawick2010numerical,
title={Numerical Simulation of the Complex Ginzburg-Landau Equation on GPUs with CUDA},
author={Hawick, KA and Playne, DP},
year={2010},
organization={Massey University, Tech. Rep. CSTN-070, January 2010, to appear in Proc. IASTED International Conference on Parallel and Distributed Computing and Networks (PDCN), 14-16 Feb. 2011, in Innsbruck, Austria}
}
The Time Dependent Ginzburg Landau(TDGL) equation models a complex scalar field and is used to study a variety of different physical systems and exhibits phase transitional behaviours that necessitate study using numerical simulation methods. We employ fast data-parallel simulation algorithms on Graphical Processing Units (GPUs) and report on performance data and stability tradeoffs from using various implementations of both 32-bit and 64-bit complex numbers. Using NVIDIA’s Compute Unified Device Architecture (CUDA) programming language running on a GTX480 GPU, we are able to simulate the TDGL with relatively large simulation system sizes of 256^3 cells and we discuss the relative computational tradeoffs between numerical accuracy and stability using different methods as well as different data precisions.
February 14, 2011 by hgpu
Your response
You must be logged in to post a comment.