Timezone: »
Training recurrent neural networks (RNNs) on low-dimensional tasks has been widely used to model functional biological networks. However, the solutions found by learning and the effect of initial connectivity are not well understood. Here, we examine RNNs trained using gradient descent on different tasks inspired by the neuroscience literature. We find that the changes in recurrent connectivity can be described by low-rank matrices. This observation holds even in the presence of random initial connectivity, although this initial connectivity has full rank and significantly accelerates training. To understand the origin of these observations, we turn to an analytically tractable setting: training a linear RNN on a simpler task. We show how the low-dimensional task structure leads to low-rank changes to connectivity, and how random initial connectivity facilitates learning. Altogether, our study opens a new perspective to understand learning in RNNs in light of low-rank connectivity changes and the synergistic role of random initialization.
Author Information
Friedrich Schuessler (Technion)
Francesca Mastrogiuseppe (UCL)
Alexis Dubreuil (ENS)
Srdjan Ostojic (Ecole Normale Superieure)
Omri Barak (Technion - Israeli institute of technology)
Related Events (a corresponding poster, oral, or spotlight)
-
2020 Oral: The interplay between randomness and structure during learning in RNNs »
Wed. Dec 9th 02:30 -- 02:45 PM Room Orals & Spotlights: Deep Learning
More from the Same Authors
-
2023 Poster: The Simplicity Bias in Multi-Task RNNs: Shared Attractors, Reuse of Dynamics, and Geometric Representation »
Elia Turner · Omri Barak -
2022 Poster: Extracting computational mechanisms from neural data using low-rank RNNs »
Adrian Valente · Jonathan Pillow · Srdjan Ostojic -
2021 Poster: Charting and Navigating the Space of Solutions for Recurrent Neural Networks »
Elia Turner · Kabir V Dabholkar · Omri Barak