Segmented-Memory Recurrent Neural Networks

被引:16
|
作者
Chen, Jinmiao [1 ]
Chaudhari, Narendra S. [1 ]
机构
[1] Nanyang Technol Univ, Sch Comp Engn, Singapore 639798, Singapore
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2009年 / 20卷 / 08期
关键词
Gradient descent; information latching; long-term dependencies; recurrent neural networks (RNNs); segmented memory; vanishing gradient; PROTEIN SECONDARY STRUCTURE; LONG-TERM DEPENDENCIES; STRUCTURE PREDICTION; GRADIENT-DESCENT; REHEARSAL; ALGORITHM; STATE;
D O I
10.1109/TNN.2009.2022980
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Conventional recurrent neural networks (RNNs) have difficulties in learning long-term dependencies. To tackle this problem, we propose an architecture called segmented-memory recurrent neural network (SMRNN). A symbolic sequence is broken into segments and then presented as inputs to the SMRNN one symbol per cycle. The SMRNN uses separate internal states to store symbol-level context, as well as segment-level context. The symbol-level context is updated for each symbol presented for input. The segment-level context is updated after each segment. The SMRNN is trained using an extended real-time recurrent learning algorithm. We test the performance of SMRNN on the information latching problem, the "two-sequence problem" and the problem of protein secondary structure (PSS) prediction. Our implementation results indicate that SMRNN performs better on long-term dependency problems than conventional RNNs. Besides, we also theoretically analyze how the segmented memory of SMRNN helps learning long-term temporal dependencies and study the impact of the segment length.
引用
收藏
页码:1267 / 1280
页数:14
相关论文
共 50 条
  • [21] Experimental Evaluation of Memory Capacity of Recurrent Neural Networks
    Kolesau, Aliaksei
    Sesok, Dmitrij
    Goranin, Nikolaj
    Rybokas, Mindaugas
    BALTIC JOURNAL OF MODERN COMPUTING, 2019, 7 (01): : 138 - 150
  • [22] Memory Analysis for Memristors and Memristive Recurrent Neural Networks
    Gang Bao
    Yide Zhang
    Zhigang Zeng
    IEEE/CAAJournalofAutomaticaSinica, 2020, 7 (01) : 96 - 105
  • [23] Memory in linear recurrent neural networks in continuous time
    Hermans, Michiel
    Schrauwen, Benjamin
    NEURAL NETWORKS, 2010, 23 (03) : 341 - 355
  • [24] Encoding-based memory for recurrent neural networks
    Carta, Antonio
    Sperduti, Alessandro
    Bacciu, Davide
    NEUROCOMPUTING, 2021, 456 (456) : 407 - 420
  • [25] Memory analysis for memristors and memristive recurrent neural networks
    Bao, Gang
    Zhang, Yide
    Zeng, Zhigang
    IEEE-CAA JOURNAL OF AUTOMATICA SINICA, 2020, 7 (01) : 96 - 105
  • [26] State-Frequency Memory Recurrent Neural Networks
    Hu, Hao
    Qi, Guo-Jun
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [27] Associative memory by recurrent neural networks with delay elements
    Miyoshi, S
    Yanai, HF
    Okada, M
    ICONIP'02: PROCEEDINGS OF THE 9TH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING: COMPUTATIONAL INTELLIGENCE FOR THE E-AGE, 2002, : 70 - 74
  • [28] Associative memory by recurrent neural networks with delay elements
    Miyoshi, S
    Yanai, HF
    Okada, M
    NEURAL NETWORKS, 2004, 17 (01) : 55 - 63
  • [29] Neural Mechanisms of Working Memory Accuracy Revealed by Recurrent Neural Networks
    Xie, Yuanqi
    Liu, Yichen Henry
    Constantinidis, Christos
    Zhou, Xin
    FRONTIERS IN SYSTEMS NEUROSCIENCE, 2022, 16
  • [30] FARM: A Flexible Accelerator for Recurrent and Memory Augmented Neural Networks
    Nagadastagiri Challapalle
    Sahithi Rampalli
    Nicholas Jao
    Akshaykrishna Ramanathan
    John Sampson
    Vijaykrishnan Narayanan
    Journal of Signal Processing Systems, 2020, 92 : 1247 - 1261