Segmented-Memory Recurrent Neural Networks

被引:16
|
作者
Chen, Jinmiao [1 ]
Chaudhari, Narendra S. [1 ]
机构
[1] Nanyang Technol Univ, Sch Comp Engn, Singapore 639798, Singapore
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2009年 / 20卷 / 08期
关键词
Gradient descent; information latching; long-term dependencies; recurrent neural networks (RNNs); segmented memory; vanishing gradient; PROTEIN SECONDARY STRUCTURE; LONG-TERM DEPENDENCIES; STRUCTURE PREDICTION; GRADIENT-DESCENT; REHEARSAL; ALGORITHM; STATE;
D O I
10.1109/TNN.2009.2022980
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Conventional recurrent neural networks (RNNs) have difficulties in learning long-term dependencies. To tackle this problem, we propose an architecture called segmented-memory recurrent neural network (SMRNN). A symbolic sequence is broken into segments and then presented as inputs to the SMRNN one symbol per cycle. The SMRNN uses separate internal states to store symbol-level context, as well as segment-level context. The symbol-level context is updated for each symbol presented for input. The segment-level context is updated after each segment. The SMRNN is trained using an extended real-time recurrent learning algorithm. We test the performance of SMRNN on the information latching problem, the "two-sequence problem" and the problem of protein secondary structure (PSS) prediction. Our implementation results indicate that SMRNN performs better on long-term dependency problems than conventional RNNs. Besides, we also theoretically analyze how the segmented memory of SMRNN helps learning long-term temporal dependencies and study the impact of the segment length.
引用
收藏
页码:1267 / 1280
页数:14
相关论文
共 50 条
  • [41] Accelerating Recurrent Neural Networks: A Memory-Efficient Approach
    Wang, Zhisheng
    Lin, Jun
    Wang, Zhongfeng
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2017, 25 (10) : 2763 - 2775
  • [42] Using Stigmergy as a Computational Memory in the Design of Recurrent Neural Networks
    Galatolo, Federico A.
    Cimino, Mario G. C. A.
    Vaglini, Gigliola
    ICPRAM: PROCEEDINGS OF THE 8TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION APPLICATIONS AND METHODS, 2019, : 830 - 836
  • [43] On the Duration, Addressability, and Capacity of Memory-Augmented Recurrent Neural Networks
    Quan, Zhibin
    Gao, Zhiqiang
    Zeng, Weili
    Li, Xuelian
    Zhu, Man
    IEEE ACCESS, 2018, 6 : 12462 - 12472
  • [44] Evaluation of memory capacity of spin torque oscillator for recurrent neural networks
    Tsunegi, Sumito
    Taniguchi, Tomohiro
    Miwa, Shinji
    Nakajima, Kohei
    Yakushiji, Kay
    Fukushima, Akio
    Yuasa, Shinji
    Kubota, Hitoshi
    JAPANESE JOURNAL OF APPLIED PHYSICS, 2018, 57 (12)
  • [45] Recurrent neural networks with small weights implement definite memory machines
    Hammer, B
    Tino, P
    NEURAL COMPUTATION, 2003, 15 (08) : 1897 - 1929
  • [46] Using recurrent neural networks to optimize dynamical decoupling for quantum memory
    August, Moritz
    Ni, Xiaotong
    PHYSICAL REVIEW A, 2017, 95 (01)
  • [47] Memory consolidation and improvement by synaptic tagging and capture in recurrent neural networks
    Jannik Luboeinski
    Christian Tetzlaff
    Communications Biology, 4
  • [48] EXPERIMENTAL STUDIES OF MEMORY SURFACES AND LEARNING SURFACES IN RECURRENT NEURAL NETWORKS
    WATANABE, T
    UCHIKAWA, Y
    GOUHARA, K
    SYSTEMS AND COMPUTERS IN JAPAN, 1994, 25 (08) : 27 - 39
  • [49] Memory consolidation and improvement by synaptic tagging and capture in recurrent neural networks
    Luboeinski, Jannik
    Tetzlaff, Christian
    COMMUNICATIONS BIOLOGY, 2021, 4 (01)
  • [50] Associative Memory Synthesis Based on Region Attractive Recurrent Neural Networks
    Bao, Gang
    Gong, Shunqi
    Zhou, Xue
    NEURAL PROCESSING LETTERS, 2022, 54 (05) : 4593 - 4607