Opening the Black Box: Low-Dimensional Dynamics in High-Dimensional Recurrent Neural Networks

被引:235
|
作者
Sussillo, David [1 ]
Barak, Omri [2 ]
机构
[1] Stanford Univ, Dept Elect Engn, Neurosci Program, Stanford, CA 94305 USA
[2] Columbia Univ Coll Phys & Surg, Dept Neurosci, New York, NY 10032 USA
关键词
CIRCUITS; SYSTEMS; STATES; MODEL;
D O I
10.1162/NECO_a_00409
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recurrent neural networks (RNNs) are useful tools for learning nonlinear relationships between time-varying inputs and outputs with complex temporal dependencies. Recently developed algorithms have been successful at training RNNs to perform a wide variety of tasks, but the resulting networks have been treated as black boxes: their mechanism of operation remains unknown. Here we explore the hypothesis that fixed points, both stable and unstable, and the linearized dynamics around them, can reveal crucial aspects of how RNNs implement their computations. Further, we explore the utility of linearization in areas of phase space that are not true fixed points but merely points of very slow movement. We present a simple optimization technique that is applied to trained RNNs to find the fixed and slow points of their dynamics. Linearization around these slow regions can be used to explore, or reverse-engineer, the behavior of the RNN. We describe the technique, illustrate it using simple examples, and finally showcase it on three high-dimensional RNN examples: a 3-bit flip-flop device, an input-dependent sine wave generator, and a two-point moving average. In all cases, the mechanisms of trained networks could be inferred from the sets of fixed and slow points and the linearized dynamics around them.
引用
收藏
页码:626 / 649
页数:24
相关论文
共 50 条
  • [1] Low-dimensional criticality embedded in high-dimensional awake brain dynamics
    Fontenele, Antonio J.
    Sooter, J. Samuel
    Norman, V. Kindler
    Gautam, Shree Hari
    Shew, Woodrow L.
    [J]. SCIENCE ADVANCES, 2024, 10 (17):
  • [2] Mapping Low-Dimensional Dynamics to High-Dimensional Neural Activity: A Derivation of the Ring Model From the Neural Engineering Framework
    Barak, Omri
    Romani, Sandro
    [J]. NEURAL COMPUTATION, 2021, 33 (03) : 827 - 852
  • [3] Low-Dimensional Shaping for High-Dimensional Lattice Codes
    Ferdinand, Nuwan S.
    Kurkoski, Brian M.
    Nokleby, Matthew
    Aazhang, Behnaam
    [J]. IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2016, 15 (11) : 7405 - 7418
  • [4] High-dimensional dynamics of generalization error in neural networks
    Advani, Madhu S.
    Saxe, Andrew M.
    Sompolinsky, Haim
    [J]. NEURAL NETWORKS, 2020, 132 : 428 - 446
  • [5] Robust Regression Estimation Based on Low-Dimensional Recurrent Neural Networks
    Xia, Youshen
    Wang, Jun
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (12) : 5935 - 5946
  • [6] Low-Dimensional Manifolds Support Multiplexed Integrations in Recurrent Neural Networks
    Fanthomme, Arnaud
    Monasson, Remi
    [J]. NEURAL COMPUTATION, 2021, 33 (04) : 1063 - 1112
  • [7] Low-Dimensional Approximations of High-Dimensional Asset Price Models
    Redmann, Martin
    Bayer, Christian
    Goyal, Pawan
    [J]. SIAM JOURNAL ON FINANCIAL MATHEMATICS, 2021, 12 (01): : 1 - 28
  • [8] Sampling High-Dimensional Bandlimited Fields on Low-Dimensional Manifolds
    Unnikrishnan, Jayakrishnan
    Vetterli, Martin
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2013, 59 (04) : 2103 - 2127
  • [9] ON LOW-DIMENSIONAL FACES THAT HIGH-DIMENSIONAL POLYTOPES MUST HAVE
    KALAI, G
    [J]. COMBINATORICA, 1990, 10 (03) : 271 - 280
  • [10] Low-dimensional paradigms for high-dimensional hetero-chaos
    Saiki, Yoshitaka
    Sanjuan, Miguel A. F.
    Yorke, James A.
    [J]. CHAOS, 2018, 28 (10)