Recurrent neural networks (RNNs) trained on low-dimensional tasks have been widely used to model functional biological networks. However, the solutions found by learning and the effect of initial connectivity are not well understood. Here, we examine RNNs trained using gradient descent on different tasks inspired by the neuroscience literature. We find that the changes in recurrent connectivity can be described by low-rank matrices, despite the unconstrained nature of the learning algorithm. To identify the origin of the low-rank structure, we turn to an analytically tractable setting: training a linear RNN on a simplified task. We show how the low-dimensional task structure leads to low-rank changes to connectivity. This low-rank structure allows us to explain and quantify the phenomenon of accelerated learning in the presence of random initial connectivity. Altogether, our study opens a new perspective to understanding trained RNNs in terms of both the learning process and the resulting network structure.

The interplay between randomness and structure during learning in RNNs / Schuessler, Friedrich; Mastrogiuseppe, Francesca; Dubreuil, Alexis; Ostojic, Srdjan; Barak, Omri. - 2020-December:(2020). ( 34th Conference on Neural Information Processing Systems, NeurIPS 2020 Virtual, Online 6 December 2020 - 12 December 2020).

The interplay between randomness and structure during learning in RNNs

Mastrogiuseppe, Francesca;
2020-01-01

Abstract

Recurrent neural networks (RNNs) trained on low-dimensional tasks have been widely used to model functional biological networks. However, the solutions found by learning and the effect of initial connectivity are not well understood. Here, we examine RNNs trained using gradient descent on different tasks inspired by the neuroscience literature. We find that the changes in recurrent connectivity can be described by low-rank matrices, despite the unconstrained nature of the learning algorithm. To identify the origin of the low-rank structure, we turn to an analytically tractable setting: training a linear RNN on a simplified task. We show how the low-dimensional task structure leads to low-rank changes to connectivity. This low-rank structure allows us to explain and quantify the phenomenon of accelerated learning in the presence of random initial connectivity. Altogether, our study opens a new perspective to understanding trained RNNs in terms of both the learning process and the resulting network structure.
2020
NEURIPS 2020. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS
2020-December
https://arxiv.org/abs/2006.11036
Schuessler, Friedrich; Mastrogiuseppe, Francesca; Dubreuil, Alexis; Ostojic, Srdjan; Barak, Omri
File in questo prodotto:
File Dimensione Formato  
schuessler_neurips_2020.pdf

non disponibili

Descrizione: pdf editoriale
Tipologia: Versione Editoriale (PDF)
Licenza: Non specificato
Dimensione 473.33 kB
Formato Adobe PDF
473.33 kB Adobe PDF   Visualizza/Apri   Richiedi una copia

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/20.500.11767/148439
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 37
  • ???jsp.display-item.citation.isi??? 34
social impact