Poster (After Call) FZJ-2024-02719

http://join2-wiki.gsi.de/foswiki/pub/Main/Artwork/join2_logo100x88.png
Continual learning using dendritic modulations on view-invariant feedforward weights

 ;  ;

2024

Computational and Systems Neuroscience 2024, COSYNE 2024, LisbonLisbon, Portugal, 29 Feb 2024 - 3 Mar 20242024-02-292024-03-03 [10.34734/FZJ-2024-02719]

This record in other databases:

Please use a persistent id in citations: doi:

Abstract: The brain is remarkably adept at learning from a continuous stream of data without significantlyforgetting previously learnt skills. Conventional machine learning models struggle at continual learn-ing, as weight updates that optimize the current task interfere with previously learnt tasks. A simpleremedy to catastrophic forgetting is freezing a network pretrained on a set of base tasks, and trainingtask-specific readouts on this shared trunk. However, this assumes that representations in the frozennetwork are separable under new tasks, therefore leading to sub-par performance. To continually learnon novel task data, previous methods suggest weight consolidation – preserving weights that are mostimpactful for the performance of previous tasks – and memory-based approaches – where the networkis allowed to see a subset of images from previous tasks.For biological networks, prior work showed that dendritic top-down modulations provide a powerfulmechanism to learn novel tasks while initial feedforward weights solely extract generic view-invariantfeatures. Therefore, we propose a continual learner that optimizes the feedforward weights towardsview-invariant representations while training task-specific modulations towards separable class clus-ters. In a task-incremental setting, we train feedforward weights using a self-supervised algorithm,while training the task-specific modulations and readouts in a supervised fashion, both exclusivelythrough current-task data. We show that this simple approach avoids catastrophic forgetting of classclusters, as opposed to training the whole network in a supervised manner, while also outperforming(a) task-specific readout without modulations and (b) frozen feedforward weights. This suggests that(a) top-down modulations are necessary and sufficient to shift the representations towards separableclusters and that (b) the SSL objective learns novel features based on the newly presented objectswhile maintaining features relevant to previous tasks, without requiring specific synaptic consolidationmechanisms.


Contributing Institute(s):
  1. Neuromorphic Software Eco System (PGI-15)
Research Program(s):
  1. 5234 - Emerging NC Architectures (POF4-523) (POF4-523)
  2. Functional Neural Architectures (jinm60_20190501) (jinm60_20190501)
  3. WestAI - AI Service Center West (01IS22094B) (01IS22094B)

Appears in the scientific report 2024
Database coverage:
OpenAccess
Click to display QR Code for this record

The record appears in these collections:
Dokumenttypen > Präsentationen > Poster
Institutssammlungen > PGI > PGI-15
Workflowsammlungen > Öffentliche Einträge
Publikationsdatenbank
Open Access

 Datensatz erzeugt am 2024-04-12, letzte Änderung am 2024-11-28


OpenAccess:
Volltext herunterladen PDF
Dieses Dokument bewerten:

Rate this document:
1
2
3
 
(Bisher nicht rezensiert)