Opening the Black Box: Low-Dimensional Dynamics in High-Dimensional Recurrent Neural Networks

  • David Sussillo
    Department of Electrical Engineering, Neurosciences Program, Stanford University, Stanford, CA 94305-9505, U.S.A.
  • Omri Barak
    Department of Neuroscience, Columbia University College of Physicians and Surgeons, New York, NY 10032-2695, U.S.A.

説明

<jats:p>Recurrent neural networks (RNNs) are useful tools for learning nonlinear relationships between time-varying inputs and outputs with complex temporal dependencies. Recently developed algorithms have been successful at training RNNs to perform a wide variety of tasks, but the resulting networks have been treated as black boxes: their mechanism of operation remains unknown. Here we explore the hypothesis that fixed points, both stable and unstable, and the linearized dynamics around them, can reveal crucial aspects of how RNNs implement their computations. Further, we explore the utility of linearization in areas of phase space that are not true fixed points but merely points of very slow movement. We present a simple optimization technique that is applied to trained RNNs to find the fixed and slow points of their dynamics. Linearization around these slow regions can be used to explore, or reverse-engineer, the behavior of the RNN. We describe the technique, illustrate it using simple examples, and finally showcase it on three high-dimensional RNN examples: a 3-bit flip-flop device, an input-dependent sine wave generator, and a two-point moving average. In all cases, the mechanisms of trained networks could be inferred from the sets of fixed and slow points and the linearized dynamics around them.</jats:p>

収録刊行物

被引用文献 (3)*注記

もっと見る

詳細情報 詳細情報について

問題の指摘

ページトップへ