Low-Dimensional Manifolds Support Multiplexed Integrations in Recurrent Neural Networks

Arnaud Fanthomme, Rémi Monasson

Research output: Contribution to journalLetterpeer-review

Abstract

We study the learning dynamics and the representations emerging in recurrent neural networks (RNNs) trained to integrate one or multiple temporal signals. Combining analytical and numerical investigations, we characterize the conditions under which an RNN with n neurons learns to integrate D(≪n) scalar signals of arbitrary duration. We show, for lin-ear, ReLU, and sigmoidal neurons, that the internal state lives close to a D-dimensional manifold, whose shape is related to the activation func-tion. Each neuron therefore carries, to various degrees, information about the value of all integrals. We discuss the deep analogy between our results and the concept of mixed selectivity forged by computational neurosci-entists to interpret cortical recordings.

Original languageEnglish
Pages (from-to)1063-1112
Number of pages50
JournalNeural Computation
Volume33
Issue number4
DOIs
Publication statusPublished - 26 Mar 2021
Externally publishedYes

Fingerprint

Dive into the research topics of 'Low-Dimensional Manifolds Support Multiplexed Integrations in Recurrent Neural Networks'. Together they form a unique fingerprint.

Cite this