001     1041678
005     20250505202225.0
037 _ _ |a FZJ-2025-02383
041 _ _ |a English
100 1 _ |a Bouss, Peter
|0 P:(DE-Juel1)178725
|b 0
|e Corresponding author
|u fzj
111 2 _ |a 32nd Annual Computational Neuroscience Meeting
|g CNS*2023
|c Leipzig
|d 2023-07-15 - 2023-07-19
|w Germany
245 _ _ |a Nonlinear dimensionality reduction with normalizing flows for analysis of electrophysiological recordings
260 _ _ |c 2023
336 7 _ |a Conference Paper
|0 33
|2 EndNote
336 7 _ |a INPROCEEDINGS
|2 BibTeX
336 7 _ |a conferenceObject
|2 DRIVER
336 7 _ |a CONFERENCE_POSTER
|2 ORCID
336 7 _ |a Output Types/Conference Poster
|2 DataCite
336 7 _ |a Poster
|b poster
|m poster
|0 PUB:(DE-HGF)24
|s 1746441757_14472
|2 PUB:(DE-HGF)
|x After Call
520 _ _ |a Despite the large number of active neurons in the cortex, the activity of neural populations for different brain regions is expected to live on a low-dimensional manifold [1]. Among the most common tools to estimate the mapping to this manifold, along with its dimension, are variants of principal component analysis. Although their success is undisputed, these methods still have the disadvantage of assuming that the data is well described by a Gaussian distribution; any additional features such as skewness or bimodality are neglected. Their performance when used as a generative model is therefore often poor.To fully learn the statistics of neural activity and to generate artificial samples, we use Normalizing Flows (NFs) [2, 3]. These neural networks learn a dimension-preserving estimator of the probability distribution of the data (Fig. 1: Left-hand side). They differ from generative adversarial networks (GANs) and variational autoencoders (VAEs) by their simplicity – only one bijective mapping is learned – and by their ability to compute the likelihood exactly due to tractable Jacobians at each building block.We adapt the training objective of NFs to discriminate between relevant (in manifold) and noise dimensions (out of manifold). To do this, we break the original symmetry of the latent space by enforcing maximal variance of the data to be encoded by as few dimensions as possible (Fig. 1: Right-hand side)—the same idea underlying PCA, a linear model, adapted here for nonlinear mappings. This allows us to estimate the dimensionality of the neural manifold and even to describe the underlying manifold without discarding any information, a unique feature of NFs.We prove the validity of our adaptation on artificial datasets of varying complexity generated by a hidden manifold model where the underlying dimensionality is known. We illustrate the power of our approach by reconstructing data using only a few latent NF dimensions. In this setting, we show the advantage of such a nonlinear approach over linear methods.Following this approach, we identify manifolds in EEG recordings from a dataset featuring high gamma activity. As described in [4], these recordings are obtained from 128 electrodes during four movement tasks. When plotted along the first principal components obtained by PCA, these data show for some PCs a heavy-tailed distribution. While linear models such as PCA are limited to Gaussian statistics and hence suboptimal in such a case, the nonlinearity of NFs enable to learn higher-order correlations. Moreover, by flattening out the curvature in latent space, we can better associate features with latent dimensions. Especially, we have now a reduced set of latent dimensions that explain most of the data variance.References 1. Gallego JA, Perich MG, Miller LE, Solla SA. Neural manifolds for the control of movement. Neuron. 2017 Jun 7;94(5):978–84. 2. Dinh L, Krueger D, Bengio Y. Nice: Non-linear independent components estimation. arXiv preprint arXiv:1410.8516. 2014 Oct 30. 3. Dinh L, Sohl-Dickstein J, Bengio S. Density estimation using real NVP. 5th Int. InConf. Learn. Represent. ICLR 2017. 4. Schirrmeister RT, Springenberg JT, Fiederer LD, Glasstetter M, Eggensperger K, Tangermann M, Hutter F, Burgard W, Ball T. Deep learning with convolutional neural networks for EEG decoding and visualization. Human brain mapping. 2017 Nov;38(11):5391–420.
536 _ _ |a 5232 - Computational Principles (POF4-523)
|0 G:(DE-HGF)POF4-5232
|c POF4-523
|f POF IV
|x 0
536 _ _ |a 5234 - Emerging NC Architectures (POF4-523)
|0 G:(DE-HGF)POF4-5234
|c POF4-523
|f POF IV
|x 1
536 _ _ |a GRK 2416 - GRK 2416: MultiSenses-MultiScales: Neue Ansätze zur Aufklärung neuronaler multisensorischer Integration (368482240)
|0 G:(GEPRIS)368482240
|c 368482240
|x 2
536 _ _ |a RenormalizedFlows - Transparent Deep Learning with Renormalized Flows (BMBF-01IS19077A)
|0 G:(DE-Juel-1)BMBF-01IS19077A
|c BMBF-01IS19077A
|x 3
700 1 _ |a Nestler, Sandra
|0 P:(DE-Juel1)204281
|b 1
700 1 _ |a Fischer, Kirsten
|0 P:(DE-Juel1)180150
|b 2
|u fzj
700 1 _ |a Merger, Claudia Lioba
|0 P:(DE-Juel1)184900
|b 3
700 1 _ |a Rene, Alexandre
|0 P:(DE-Juel1)178936
|b 4
700 1 _ |a Helias, Moritz
|0 P:(DE-Juel1)144806
|b 5
|u fzj
856 4 _ |u https://link.springer.com/article/10.1007/s10827-024-00871-5
909 C O |o oai:juser.fz-juelich.de:1041678
|p VDB
910 1 _ |a Forschungszentrum Jülich
|0 I:(DE-588b)5008462-8
|k FZJ
|b 0
|6 P:(DE-Juel1)178725
910 1 _ |a RWTH Aachen
|0 I:(DE-588b)36225-6
|k RWTH
|b 0
|6 P:(DE-Juel1)178725
910 1 _ |a Forschungszentrum Jülich
|0 I:(DE-588b)5008462-8
|k FZJ
|b 1
|6 P:(DE-Juel1)204281
910 1 _ |a RWTH Aachen
|0 I:(DE-588b)36225-6
|k RWTH
|b 1
|6 P:(DE-Juel1)204281
910 1 _ |a Forschungszentrum Jülich
|0 I:(DE-588b)5008462-8
|k FZJ
|b 2
|6 P:(DE-Juel1)180150
910 1 _ |a RWTH Aachen
|0 I:(DE-588b)36225-6
|k RWTH
|b 2
|6 P:(DE-Juel1)180150
910 1 _ |a Forschungszentrum Jülich
|0 I:(DE-588b)5008462-8
|k FZJ
|b 3
|6 P:(DE-Juel1)184900
910 1 _ |a RWTH Aachen
|0 I:(DE-588b)36225-6
|k RWTH
|b 3
|6 P:(DE-Juel1)184900
910 1 _ |a Forschungszentrum Jülich
|0 I:(DE-588b)5008462-8
|k FZJ
|b 4
|6 P:(DE-Juel1)178936
910 1 _ |a RWTH Aachen
|0 I:(DE-588b)36225-6
|k RWTH
|b 4
|6 P:(DE-Juel1)178936
910 1 _ |a University of Ottawa, Canada
|0 I:(DE-HGF)0
|b 4
|6 P:(DE-Juel1)178936
910 1 _ |a Forschungszentrum Jülich
|0 I:(DE-588b)5008462-8
|k FZJ
|b 5
|6 P:(DE-Juel1)144806
910 1 _ |a RWTH Aachen
|0 I:(DE-588b)36225-6
|k RWTH
|b 5
|6 P:(DE-Juel1)144806
913 1 _ |a DE-HGF
|b Key Technologies
|l Natural, Artificial and Cognitive Information Processing
|1 G:(DE-HGF)POF4-520
|0 G:(DE-HGF)POF4-523
|3 G:(DE-HGF)POF4
|2 G:(DE-HGF)POF4-500
|4 G:(DE-HGF)POF
|v Neuromorphic Computing and Network Dynamics
|9 G:(DE-HGF)POF4-5232
|x 0
913 1 _ |a DE-HGF
|b Key Technologies
|l Natural, Artificial and Cognitive Information Processing
|1 G:(DE-HGF)POF4-520
|0 G:(DE-HGF)POF4-523
|3 G:(DE-HGF)POF4
|2 G:(DE-HGF)POF4-500
|4 G:(DE-HGF)POF
|v Neuromorphic Computing and Network Dynamics
|9 G:(DE-HGF)POF4-5234
|x 1
920 _ _ |l yes
920 1 _ |0 I:(DE-Juel1)INM-6-20090406
|k INM-6
|l Computational and Systems Neuroscience
|x 0
920 1 _ |0 I:(DE-Juel1)IAS-6-20130828
|k IAS-6
|l Computational and Systems Neuroscience
|x 1
980 _ _ |a poster
980 _ _ |a VDB
980 _ _ |a I:(DE-Juel1)INM-6-20090406
980 _ _ |a I:(DE-Juel1)IAS-6-20130828
980 _ _ |a UNRESTRICTED


LibraryCollectionCLSMajorCLSMinorLanguageAuthor
Marc 21