Preprint FZJ-2022-01560

http://join2-wiki.gsi.de/foswiki/pub/Main/Artwork/join2_logo100x88.png
Routing brain traffic through the von Neumann bottleneck: Efficient cache usage in spiking neural network simulation code on general purpose computers

 ;  ;  ;  ;  ;

2021
arXiv

arXiv () [10.48550/ARXIV.2109.12855]

This record in other databases:

Please use a persistent id in citations:   doi:

Abstract: Simulation is a third pillar next to experiment and theory in the study of complex dynamic systems such as biological neural networks. Contemporary brain-scale networks correspond to directed graphs of a few million nodes, each with an in-degree and out-degree of several thousands of edges, where nodes and edges correspond to the fundamental biological units, neurons and synapses, respectively. When considering a random graph, each node's edges are distributed across thousands of parallel processes. The activity in neuronal networks is also sparse. Each neuron occasionally transmits a brief signal, called spike, via its outgoing synapses to the corresponding target neurons. This spatial and temporal sparsity represents an inherent bottleneck for simulations on conventional computers: Fundamentally irregular memory-access patterns cause poor cache utilization. Using an established neuronal network simulation code as a reference implementation, we investigate how common techniques to recover cache performance such as software-induced prefetching and software pipelining can benefit a real-world application. The algorithmic changes reduce simulation time by up to 50%. The study exemplifies that many-core systems assigned with an intrinsically parallel computational problem can overcome the von Neumann bottleneck of conventional computer architectures.

Keyword(s): Distributed, Parallel, and Cluster Computing (cs.DC) ; FOS: Computer and information sciences


Contributing Institute(s):
  1. Computational and Systems Neuroscience (INM-6)
  2. Theoretical Neuroscience (IAS-6)
  3. Jara-Institut Brain structure-function relationships (INM-10)
Research Program(s):
  1. 5234 - Emerging NC Architectures (POF4-523) (POF4-523)
  2. HBP SGA2 - Human Brain Project Specific Grant Agreement 2 (785907) (785907)
  3. HBP SGA3 - Human Brain Project Specific Grant Agreement 3 (945539) (945539)
  4. DEEP-EST - DEEP - Extreme Scale Technologies (754304) (754304)
  5. ACA - Advanced Computing Architectures (SO-092) (SO-092)
  6. GRK 2416:  MultiSenses-MultiScales: Novel approaches to decipher neural processing in multisensory integration (368482240) (368482240)
  7. ATMLPP - ATML Parallel Performance (ATMLPP) (ATMLPP)

Appears in the scientific report 2022
Database coverage:
OpenAccess
Click to display QR Code for this record

The record appears in these collections:
Institutssammlungen > INM > INM-10
Institutssammlungen > IAS > IAS-6
Institutssammlungen > INM > INM-6
Dokumenttypen > Berichte > Vorabdrucke
Workflowsammlungen > Öffentliche Einträge
Publikationsdatenbank
Open Access

 Datensatz erzeugt am 2022-03-11, letzte Änderung am 2025-03-14


Dieses Dokument bewerten:

Rate this document:
1
2
3
 
(Bisher nicht rezensiert)