Approximation and Optimization Theory for Linear Continuous-Time Recurrent Neural Networks

被引:0
|
作者
Li, Zhong [1 ]
Han, Jiequn [2 ]
E, Weinan [2 ,3 ]
Li, Qianxiao [4 ]
机构
[1] Peking Univ, Sch Math Sci, Beijing 100080, Peoples R China
[2] Princeton Univ, Dept Math, Princeton, NJ 08544 USA
[3] Princeton Univ, PACM, Princeton, NJ 08544 USA
[4] Natl Univ Singapore, Dept Math, Singapore 119076, Singapore
关键词
recurrent neural networks; dynamical systems; approximation; optimization; curse of memory; DYNAMICAL-SYSTEMS; GRADIENT DESCENT; OPERATORS; MEMORY; SUMS;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We perform a systematic study of the approximation properties and optimization dynamics of recurrent neural networks (RNNs) when applied to learn input-output relationships in temporal data. We consider the simple but representative setting of using continuous-time linear RNNs to learn from data generated by linear relationships. On the approximation side, we prove a direct and an inverse approximation theorem of linear functionals using RNNs, which reveal the intricate connections between memory structures in the target and the corresponding approximation efficiency. In particular, we show that temporal relationships can be effectively approximated by RNNs if and only if the former possesses sufficient memory decay. On the optimization front, we perform detailed analysis of the optimization dynamics, including a precise understanding of the difficulty that may arise in learning relationships with long-term memory. The term "curse of memory" is coined to describe the uncovered phenomena, akin to the "curse of dimension" that plagues high dimensional function approximation. These results form a relatively complete picture of the interaction of memory and recurrent structures in the linear dynamical setting.
引用
收藏
页码:1 / 85
页数:85
相关论文
共 50 条
  • [1] Approximation and Optimization Theory for Linear Continuous-Time Recurrent Neural Networks
    Li, Zhong
    Han, Jiequn
    Weinan, E.
    Li, Qianxiao
    [J]. Journal of Machine Learning Research, 2022, 23
  • [2] Self-Optimization in Continuous-Time Recurrent Neural Networks
    Zarco, Mario
    Froese, Tom
    [J]. FRONTIERS IN ROBOTICS AND AI, 2018, 5
  • [3] APPROXIMATION OF DYNAMICAL-SYSTEMS BY CONTINUOUS-TIME RECURRENT NEURAL NETWORKS
    FUNAHASHI, K
    NAKAMURA, Y
    [J]. NEURAL NETWORKS, 1993, 6 (06) : 801 - 806
  • [4] Approximation of dynamical time-variant systems by continuous-time recurrent neural networks
    Li, XD
    Ho, JKL
    Chow, TWS
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2005, 52 (10) : 656 - 660
  • [5] Complete controllability of continuous-time recurrent neural networks
    Sontag, E
    Sussmann, H
    [J]. SYSTEMS & CONTROL LETTERS, 1997, 30 (04) : 177 - 183
  • [6] A learning result for continuous-time recurrent neural networks
    Sontag, Eduardo D.
    [J]. Systems and Control Letters, 1998, 34 (03): : 151 - 158
  • [7] ON THE DYNAMICS OF SMALL CONTINUOUS-TIME RECURRENT NEURAL NETWORKS
    BEER, RD
    [J]. ADAPTIVE BEHAVIOR, 1995, 3 (04) : 469 - 509
  • [8] A learning result for continuous-time recurrent neural networks
    Sontag, ED
    [J]. SYSTEMS & CONTROL LETTERS, 1998, 34 (03) : 151 - 158
  • [9] Noisy recurrent neural networks: The continuous-time case
    Das, S
    Olurotimi, O
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 1998, 9 (05): : 913 - 936
  • [10] Global dissipativity of continuous-time recurrent neural networks with time delay
    Liao, XX
    Wang, J
    [J]. PHYSICAL REVIEW E, 2003, 68 (01):