001     1050452
005     20260220104305.0
024 7 _ |a 10.48550/ARXIV.2507.06079
|2 doi
024 7 _ |a 10.34734/FZJ-2026-00222
|2 datacite_doi
037 _ _ |a FZJ-2026-00222
100 1 _ |a Siegel, Sebastian
|0 P:(DE-Juel1)174486
|b 0
|e Corresponding author
245 _ _ |a QS4D: Quantization-aware training for efficient hardware deployment of structured state-space sequential models
260 _ _ |c 2025
|b arXiv
336 7 _ |a Preprint
|b preprint
|m preprint
|0 PUB:(DE-HGF)25
|s 1768997506_10779
|2 PUB:(DE-HGF)
336 7 _ |a WORKING_PAPER
|2 ORCID
336 7 _ |a Electronic Article
|0 28
|2 EndNote
336 7 _ |a preprint
|2 DRIVER
336 7 _ |a ARTICLE
|2 BibTeX
336 7 _ |a Output Types/Working Paper
|2 DataCite
520 _ _ |a Structured State Space models (SSM) have recently emerged as a new class of deep learning models, particularly well-suited for processing long sequences. Their constant memory footprint, in contrast to the linearly scaling memory demands of Transformers, makes them attractive candidates for deployment on resource-constrained edge-computing devices. While recent works have explored the effect of quantization-aware training (QAT) on SSMs, they typically do not address its implications for specialized edge hardware, for example, analog in-memory computing (AIMC) chips. In this work, we demonstrate that QAT can significantly reduce the complexity of SSMs by up to two orders of magnitude across various performance metrics. We analyze the relation between model size and numerical precision, and show that QAT enhances robustness to analog noise and enables structural pruning. Finally, we integrate these techniques to deploy SSMs on a memristive analog in-memory computing substrate and highlight the resulting benefits in terms of computational efficiency.
536 _ _ |a 5234 - Emerging NC Architectures (POF4-523)
|0 G:(DE-HGF)POF4-5234
|c POF4-523
|f POF IV
|x 0
536 _ _ |a BMBF 03ZU1106CB - NeuroSys: Algorithm-Hardware Co-Design (Projekt C) - B (BMBF-03ZU1106CB)
|0 G:(DE-Juel1)BMBF-03ZU1106CB
|c BMBF-03ZU1106CB
|x 1
588 _ _ |a Dataset connected to DataCite
650 _ 7 |a Machine Learning (cs.LG)
|2 Other
650 _ 7 |a Artificial Intelligence (cs.AI)
|2 Other
650 _ 7 |a FOS: Computer and information sciences
|2 Other
700 1 _ |a Yang, Ming-Jay
|0 P:(DE-Juel1)192385
|b 1
|u fzj
700 1 _ |a Bouhadjar, Younes
|0 P:(DE-Juel1)176778
|b 2
|u fzj
700 1 _ |a Fabre, Maxime
|0 P:(DE-Juel1)201205
|b 3
|u fzj
700 1 _ |a Neftci, Emre
|0 P:(DE-Juel1)188273
|b 4
|u fzj
700 1 _ |a Strachan, John Paul
|0 P:(DE-Juel1)188145
|b 5
|u fzj
773 _ _ |a 10.48550/ARXIV.2507.06079
856 4 _ |u https://juser.fz-juelich.de/record/1050452/files/2507.06079v1.pdf
|y OpenAccess
909 C O |o oai:juser.fz-juelich.de:1050452
|p openaire
|p open_access
|p VDB
|p driver
|p dnbdelivery
910 1 _ |a Forschungszentrum Jülich
|0 I:(DE-588b)5008462-8
|k FZJ
|b 0
|6 P:(DE-Juel1)174486
910 1 _ |a Forschungszentrum Jülich
|0 I:(DE-588b)5008462-8
|k FZJ
|b 1
|6 P:(DE-Juel1)192385
910 1 _ |a Forschungszentrum Jülich
|0 I:(DE-588b)5008462-8
|k FZJ
|b 2
|6 P:(DE-Juel1)176778
910 1 _ |a Forschungszentrum Jülich
|0 I:(DE-588b)5008462-8
|k FZJ
|b 3
|6 P:(DE-Juel1)201205
910 1 _ |a Forschungszentrum Jülich
|0 I:(DE-588b)5008462-8
|k FZJ
|b 4
|6 P:(DE-Juel1)188273
910 1 _ |a Forschungszentrum Jülich
|0 I:(DE-588b)5008462-8
|k FZJ
|b 5
|6 P:(DE-Juel1)188145
913 1 _ |a DE-HGF
|b Key Technologies
|l Natural, Artificial and Cognitive Information Processing
|1 G:(DE-HGF)POF4-520
|0 G:(DE-HGF)POF4-523
|3 G:(DE-HGF)POF4
|2 G:(DE-HGF)POF4-500
|4 G:(DE-HGF)POF
|v Neuromorphic Computing and Network Dynamics
|9 G:(DE-HGF)POF4-5234
|x 0
914 1 _ |y 2025
915 _ _ |a OpenAccess
|0 StatID:(DE-HGF)0510
|2 StatID
920 1 _ |0 I:(DE-Juel1)PGI-14-20210412
|k PGI-14
|l Neuromorphic Compute Nodes
|x 0
920 1 _ |0 I:(DE-Juel1)PGI-15-20210701
|k PGI-15
|l Neuromorphic Software Eco System
|x 1
980 1 _ |a FullTexts
980 _ _ |a preprint
980 _ _ |a VDB
980 _ _ |a UNRESTRICTED
980 _ _ |a I:(DE-Juel1)PGI-14-20210412
980 _ _ |a I:(DE-Juel1)PGI-15-20210701


LibraryCollectionCLSMajorCLSMinorLanguageAuthor
Marc 21