TY - CONF
AU - Villamar, Jose
AU - Tiddia, Gianmarco
AU - Sergi, Luca
AU - Babu, Pooja
AU - Pontisso, Luca
AU - Simula, Francesco
AU - Lonardo, Alessandro
AU - Pastorelli, Elena
AU - Paolucci, Pier Stanislao
AU - Golosio, Bruno
AU - Senk, Johanna
TI - NEST GPU simulations scale up to networks of billions of spiking neurons and trillions of synapses
PB - RWTH Aachen
M1 - FZJ-2025-03154
PY - 2025
AB - Efficient simulation of large-scale spiking neuronal networks is important for neuroscientific research, and both the simulation speed and the time it takes to instantiate the network in computer memory are key factors. NEST GPU is a GPU-based simulator under the NEST Initiative written in CUDA-C++ that demonstrates high simulation speeds with models of various network sizes on single-GPU and multi-GPU systems [1,2,3]. On the path toward models of the whole brain, neuroscientists show an increasing interest in studying networks that are larger by several orders of magnitude. Here, we show the performance of our simulation technology with a scalable network model across multiple network sizes approaching human cortex magnitudes.For this, we propose a novel method to efficiently instantiate large networks on multiple GPUs in parallel. Our approach relies on the deterministic initial state of pseudo-random number generators (PRNGs). While requiring synchronization of network construction directives between MPI processes and a small memory overhead, this approach enables dynamical neuron creation and connection at runtime. The method is evaluated through a two-population recurrently connected network model designed for benchmarking an arbitrary number of GPUs while maintaining first-order network statistics across scales.The benchmarking model was tested during an exclusive reservation of the LEONARDO Booster cluster. While keeping constant the number of neurons and incoming synapses to each neuron per GPU, we performed several simulation runs exploiting in parallel from 400 to 12,000 (full system) GPUs. Each GPU device contained approximately 281 thousand neurons and 3.1 billion synapses. Our results show network construction times of less than a second using the full system and stable dynamics across scales. At full system scale, the network model was composed of approximately 3.37 billion neurons and 37.96 trillion synapses (~25% human cortex).To conclude, our novel approach enabled network model instantiation of magnitudes nearing human cortex scale while keeping fast construction times, on average of 0.5s across trials. The stability of dynamics and performance across scales obtained in our model is a proof of feasibility paving the way for biologically more plausible and detailed brain scale models. [1] https://doi.org/10.3389/fncom.2021.627620 . [2] https://doi.org/10.3389/fninf.2022.883333 . [3] https://doi.org/10.3390/app13179598
T2 - 34th Annual Computational Neuroscience Meeting
CY - 5 Jul 2025 - 9 Jul 2025, Florence (Italy)
Y2 - 5 Jul 2025 - 9 Jul 2025
M2 - Florence, Italy
LB - PUB:(DE-HGF)24
DO - DOI:10.34734/FZJ-2025-03154
UR - https://juser.fz-juelich.de/record/1044380
ER -