A Massively Parallel Adaptive Fast Multipole Method on Heterogeneous Architectures
Lawrence Livermore National Laboratory, Livermore, CA
Communications of the ACM 55, 2012
@article{Lashuk:2012:MPA:2160718.2160740,
author={Lashuk, Ilya and Chandramowlishwaran, Aparna and Langston, Harper and Nguyen, Tuan-Anh and Sampath, Rahul and Shringarpure, Aashay and Vuduc, Richard and Ying, Lexing and Zorin, Denis and Biros, George},
title={A massively parallel adaptive fast multipole method on heterogeneous architectures},
journal={Commun. ACM},
issue_date={May 2012},
volume={55},
number={5},
month={may},
year={2012},
issn={0001-0782},
pages={101–109},
numpages={9},
url={http://doi.acm.org/10.1145/2160718.2160740},
doi={10.1145/2160718.2160740},
acmid={2160740},
publisher={ACM},
address={New York, NY, USA}
}
We describe a parallel fast multipole method (FMM) for highly nonuniform distributions of particles. We employ both distributed memory parallelism (via MPI) and shared memory parallelism (via OpenMP and GPU acceleration) to rapidly evaluate two-body nonoscillatory potentials in three dimensions on heterogeneous high performance computing architectures. We have performed scalability tests with up to 30 billion particles on 196,608 cores on the AMD/CRAY-based Jaguar system at ORNL. On a GPU-enabled system (NSF’s Keeneland at Georgia Tech/ORNL), we observed 30x speedup over a single core CPU and 7x speedup over a multicore CPU implementation. By combining GPUs with MPI, we achieve less than 10ns/particle and six digits of accuracy for a run with 48 million nonuniformly distributed particles on 192 GPUs.
July 5, 2012 by hgpu