Home > Publications database > Mathematical Techniques to Reduce Memory Requirements in Deep Learning |
Conference Presentation (Other) | FZJ-2024-06888 |
2024
This record in other databases:
Please use a persistent id in citations: doi:10.34734/FZJ-2024-06888
Abstract: We present a method to substantially lower memory requirements during the training of deep neural networks, based on the GaLore (Gradient Low-Rank Projection) training framework. A rapid decay of singular values in gradient matrices permits the use of low-rank bases to encapsulate the relevant subspaces, reducing the memory requirements for storing optimizer states between iterations. A novel, rank-adaptive, GPU-optimized version of the randomized range finder algorithm is employed to exploit this property and future research directions are discussed.
![]() |
The record appears in these collections: |