Poster (After Call) FZJ-2025-03153

http://join2-wiki.gsi.de/foswiki/pub/Main/Artwork/join2_logo100x88.png
NEST GPU simulations scale up to networks of billions of spiking neurons and trillions of synapses

 ;  ;  ;  ;  ;  ;  ;  ;  ;  ;

2025

International Summer School on High-Performance Computing for Science, Industry, and Society, NaplesNaples, Italy, 16 Jun 2025 - 20 Jun 20252025-06-162025-06-20 [10.34734/FZJ-2024-05342]

This record in other databases:

Please use a persistent id in citations: doi:  doi:

Abstract: Efficient simulation of large-scale spiking neuronal networks is important for neuroscientific research, and both the simulation speed and the time it takes to instantiate the network in computer memory are key factors. NEST GPU is a GPU-based simulator under the NEST Initiative written in CUDA-C++ that demonstrates high simulation speeds with models of various network sizes on single-GPU and multi-GPU systems [1,2,3]. On the path toward models of the whole brain, neuroscientists show an increasing interest in studying networks that are larger by several orders of magnitude. Here, we show the performance of our simulation technology with a scalable network model across multiple network sizes approaching human cortex magnitudes.For this, we propose a novel method to efficiently instantiate large networks on multiple GPUs in parallel. Our approach relies on the deterministic initial state of pseudo-random number generators (PRNGs). While requiring synchronization of network construction directives between MPI processes and a small memory overhead, this approach enables dynamical neuron creation and connection at runtime. The method is evaluated through a two-population recurrently connected network model designed for benchmarking an arbitrary number of GPUs while maintaining first-order network statistics across scales.The benchmarking model was tested during an exclusive reservation of the LEONARDO Booster cluster. While keeping constant the number of neurons and incoming synapses to each neuron per GPU, we performed several simulation runs exploiting in parallel from 400 to 12,000 (full system) GPUs. Each GPU device contained approximately 281 thousand neurons and 3.1 billion synapses. Our results show network construction times of less than a second using the full system and stable dynamics across scales. At full system scale, the network model was composed of approximately 3.37 billion neurons and 37.96 trillion synapses (~25% human cortex).To conclude, our novel approach enabled network model instantiation of magnitudes nearing human cortex scale while keeping fast construction times, on average of 0.5s across trials. The stability of dynamics and performance across scales obtained in our model is a proof of feasibility paving the way for biologically more plausible and detailed brain scale models. [1] https://doi.org/10.3389/fncom.2021.627620 . [2] https://doi.org/10.3389/fninf.2022.883333 . [3] https://doi.org/10.3390/app13179598


Contributing Institute(s):
  1. Computational and Systems Neuroscience (IAS-6)
  2. Jülich Supercomputing Center (JSC)
Research Program(s):
  1. 5232 - Computational Principles (POF4-523) (POF4-523)
  2. 5235 - Digitization of Neuroscience and User-Community Building (POF4-523) (POF4-523)
  3. HiRSE - Helmholtz Platform for Research Software Engineering (HiRSE-20250220) (HiRSE-20250220)
  4. JL SMHB - Joint Lab Supercomputing and Modeling for the Human Brain (JL SMHB-2021-2027) (JL SMHB-2021-2027)

Appears in the scientific report 2025
Database coverage:
OpenAccess
Click to display QR Code for this record

The record appears in these collections:
Document types > Presentations > Poster
Institute Collections > IAS > IAS-6
Workflow collections > Public records
Institute Collections > JSC
Publications database
Open Access

 Record created 2025-07-17, last modified 2025-07-22


OpenAccess:
Download fulltext PDF
Rate this document:

Rate this document:
1
2
3
 
(Not yet reviewed)