Recurrent Neural Networks With Finite Memory Length

被引:7
|
作者
Long, Dingkun [1 ,2 ]
Zhang, Richong [1 ,2 ]
Mao, Yongyi [3 ]
机构
[1] Beihang Univ, BDBC, Beijing 100191, Peoples R China
[2] Beihang Univ, Sch Comp Sci & Engn, SKLSDE Lab, Beijing 100191, Peoples R China
[3] Univ Ottawa, Sch Elect Engn & Comp Sci, Ottawa, ON KN56N2, Canada
来源
IEEE ACCESS | 2019年 / 7卷
基金
中国国家自然科学基金;
关键词
Recurrent neural networks; memory length;
D O I
10.1109/ACCESS.2018.2890297
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The working of recurrent neural networks has not been well understood to date. The construction of such network models, hence, largely relies on heuristics and intuition. This paper formalizes the notion of "memory length" for recurrent networks and consequently discovers a generic family of recurrent networks having maximal memory lengths. Stacking such networks into multiple layers is shown to result in powerful models, including the gated convolutional networks. We show that the structure of such networks potentially enables a more principled design approach in practice and entails no gradient vanishing or exploding during back-propagation. We also present a new example in this family, termed attentive activation recurrent unit (AARU). Experimentally we demonstrate that the performance of this network family, particularly AARU, is superior to the LSTM and GRU networks.
引用
收藏
页码:12511 / 12520
页数:10
相关论文
共 50 条
  • [1] Recurrent neural networks and finite automata
    Siegelmann, HT
    COMPUTATIONAL INTELLIGENCE, 1996, 12 (04) : 567 - 574
  • [2] Minimum Description Length Recurrent Neural Networks
    Lan, Nur
    Geyer, Michal
    Chemla, Emmanuel
    Katzir, Roni
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2022, 10 : 785 - 799
  • [3] Segmented-Memory Recurrent Neural Networks
    Chen, Jinmiao
    Chaudhari, Narendra S.
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2009, 20 (08): : 1267 - 1280
  • [4] Recurrent Neural Networks and Their Memory Behavior: A Survey
    Su, Yuanhang
    Kuo, C. -C. Jay
    APSIPA TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING, 2022, 11 (01)
  • [5] Understanding and Controlling Memory in Recurrent Neural Networks
    Haviv, Doron
    Rivkind, Alexnader
    Barak, Omri
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [6] Recurrent Neural Networks With Auxiliary Memory Units
    Wang, Jianyong
    Zhang, Lei
    Guo, Quan
    Yi, Zhang
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (05) : 1652 - 1661
  • [7] Probabilistic memory capacity of recurrent neural networks
    Miyoshi, S
    Nakayama, K
    ICNN - 1996 IEEE INTERNATIONAL CONFERENCE ON NEURAL NETWORKS, VOLS. 1-4, 1996, : 1291 - 1296
  • [8] Assessing the Memory Ability of Recurrent Neural Networks
    Zhang, Cheng
    Li, Qiuchi
    Hua, Lingyu
    Song, Dawei
    ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 1658 - 1665
  • [9] On the Interpretation of Recurrent Neural Networks as Finite State Machines
    Oliva, Christian
    Lago-Fernandez, Luis F.
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: THEORETICAL NEURAL COMPUTATION, PT I, 2019, 11727 : 312 - 323
  • [10] Recurrent neural networks for musical pitch memory and classification
    Franklin, JA
    INTERNATIONAL JOURNAL ON ARTIFICIAL INTELLIGENCE TOOLS, 2005, 14 (1-2) : 329 - 342