TY  - GEN
AU  - Garcia de Gonzalo, Simon
AU  - Herten, Andreas
AU  - Hrywniak, Markus
AU  - Kraus, Jiri
AU  - Oden, Lena
TI  - Efficient Distributed GPU Programming for Exascale
M1  - FZJ-2023-05333
PY  - 2023
N1  - Github repository: https://github.com/FZJ-JSC/tutorial-multi-gpu/tree/v4.0-isc23
AB  - Over the past years, GPUs became ubiquitous in HPC installations around the world, delivering the majority of performance of some of the largest supercomputers (e.g. Summit, Sierra, JUWELS Booster). This trend continues in the recently deployed and upcoming Pre-Exascale and Exascale systems (LUMI, Leonardo; Frontier, Perlmutter): GPUs are chosen as the core computing devices to enter this next era of HPC. To take advantage of future GPU-accelerated systems with tens of thousands of devices, application developers need to have the propers skills and tools to understand, manage, and optimize distributed GPU applications. In this tutorial, participants will learn techniques to efficiently program large-scale multi-GPU systems. While programming multiple GPUs with MPI is explained in detail, also advanced tuning techniques and complementing programming models like NCCL and NVSHMEM are presented. Tools for analysis are shown and used to motivate and implement performance optimizations. The tutorial teaches fundamental concepts that apply to GPU-accelerated systems in general, taking the NVIDIA platform as an example. It is a combination of lectures and hands-on exercises, using one of Europe's fastest supercomputers, JUWELS Booster, for interactive learning and discovery.
T2  - ISC High Performance 2023
CY  - 21 May 2023 - 21 May 2023, Hamburg (Germany)
Y2  - 21 May 2023 - 21 May 2023
M2  - Hamburg, Germany
LB  - PUB:(DE-HGF)17
DO  - DOI:10.5281/ZENODO.5745504
UR  - https://juser.fz-juelich.de/record/1019363
ER  -