Advancing the distributed Multi-GPU ChASE library through algorithm optimization and NCCL library
Jülich Supercomputing Centre, Forschungszentrum Jülich GmbH, Jülich, Germany
arXiv:2309.15595 [cs.DC], (27 Sep 2023)
@misc{wu2023advancing,
title={Advancing the distributed Multi-GPU ChASE library through algorithm optimization and NCCL library},
author={Xinzhe Wu and Edoardo Di Napoli},
year={2023},
eprint={2309.15595},
archivePrefix={arXiv},
primaryClass={cs.DC}
}
As supercomputers become larger with powerful Graphics Processing Unit (GPU), traditional direct eigensolvers struggle to keep up with the hardware evolution and scale efficiently due to communication and synchronization demands. Conversely, subspace eigensolvers, like the Chebyshev Accelerated Subspace Eigensolver (ChASE), have a simpler structure and can overcome communication and synchronization bottlenecks. ChASE is a modern subspace eigensolver that uses Chebyshev polynomials to accelerate the computation of extremal eigenpairs of dense Hermitian eigenproblems. In this work we show how we have modified ChASE by rethinking its memory layout, introducing a novel parallelization scheme, switching to a more performing communication-avoiding algorithm for one of its inner modules, and substituting the MPI library by the vendor-optimized NCCL library. The resulting library can tackle dense problems with size up to N=O(106), and scales effortlessly up to the full 900 nodes — each one powered by 4×A100 NVIDIA GPUs — of the JUWELS Booster hosted at the Jülich Supercomputing Centre.
October 8, 2023 by hgpu