A Study on Catastrophic Forgetting in Deep LSTM Networks

被引:17
|
作者
Schak, Monika [1 ]
Gepperth, Alexander [1 ]
机构
[1] Univ Appl Sci Fulda, D-36037 Fulda, Germany
来源
ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: DEEP LEARNING, PT II | 2019年 / 11728卷
关键词
LSTM; Catastrophic Forgetting;
D O I
10.1007/978-3-030-30484-3_56
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a systematic study of Catastrophic Forgetting (CF), i.e., the abrupt loss of previously acquired knowledge, when retraining deep recurrent LSTM networks with new samples. CF has recently received renewed attention in the case of feed-forward DNNs, and this article is the first work that aims to rigorously establish whether deep LSTM networks are afflicted by CF as well, and to what degree. In order to test this fully, training is conducted using a wide variety of high-dimensional image-based sequence classification tasks derived from established visual classification benchmarks (MNIST, Devanagari, FashionMNIST and EMNIST). We find that the CF effect occurs universally, without exception, for deep LSTM-based sequence classifiers, regardless of the construction and provenance of sequences. This leads us to conclude that LSTMs, just like DNNs, are fully affected by CF, and that further research work needs to be conducted in order to determine how to avoid this effect (which is not a goal of this study).
引用
收藏
页码:714 / 728
页数:15
相关论文
共 50 条
  • [31] Solutions to the catastrophic forgetting problem
    Robins, A
    PROCEEDINGS OF THE TWENTIETH ANNUAL CONFERENCE OF THE COGNITIVE SCIENCE SOCIETY, 1998, : 899 - 904
  • [32] Mitigating Catastrophic Forgetting in Deep Learning in a Streaming Setting Using Historical Summary
    Dash, Sajal
    Yin, Junqi
    Shankar, Mallikarjun
    Wang, Feiyi
    Feng, Wu-chun
    PROCEEDINGS OF THE 7TH INTERNATIONAL WORKSHOP ON DATA ANALYSIS AND REDUCTION FOR BIG SCIENTIFIC DATA (DRBSD-7), 2021, : 11 - 18
  • [33] Pseudo-rehearsal: Achieving deep reinforcement learning without catastrophic forgetting
    Atkinson, Craig
    McCane, Brendan
    Szymanski, Lech
    Robins, Anthony
    NEUROCOMPUTING, 2021, 428 : 291 - 307
  • [34] Investigating Catastrophic Forgetting of Deep Learning Models Within Office 31 Dataset
    Hidayaturrahman
    Trisetyarso, Agung
    Kartowisastro, Iman Herwidiana
    Budiharto, Widodo
    IEEE ACCESS, 2024, 12 : 138501 - 138509
  • [35] Artificial Neural Variability for Deep Learning: On Overfitting, Noise Memorization, and Catastrophic Forgetting
    Xie, Zeke
    He, Fengxiang
    Fu, Shaopeng
    Sato, Issei
    Tao, Dacheng
    Sugiyama, Masashi
    NEURAL COMPUTATION, 2021, 33 (08) : 2163 - 2192
  • [36] Mixed-Privacy Forgetting in Deep Networks
    Golatkar, Aditya
    Achille, Alessandro
    Ravichandran, Avinash
    Polito, Marzia
    Soatto, Stefano
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 792 - 801
  • [37] Fast Training of Deep LSTM Networks
    Yu, Wen
    Li, Xiaoou
    Gonzalez, Jesus
    ADVANCES IN NEURAL NETWORKS - ISNN 2019, PT I, 2019, 11554 : 3 - 10
  • [38] Reducing catastrophic forgetting problem in streaming data by hybrid shark smell with jaya optimization-based deep neural networks
    Singh, Maisnam Niranjan
    Khaiyum, Samitha
    INTERNATIONAL JOURNAL OF MODELING SIMULATION AND SCIENTIFIC COMPUTING, 2022, 13 (04)
  • [39] Statistical Mechanical Analysis of Catastrophic Forgetting in Continual Learning with Teacher and Student Networks
    Asanuma, Haruka
    Takagi, Shiro
    Nagano, Yoshihiro
    Yoshida, Yuki
    Igarashi, Yasuhiko
    Okada, Masato
    JOURNAL OF THE PHYSICAL SOCIETY OF JAPAN, 2021, 90 (10)
  • [40] ADMM Consensus for Deep LSTM Networks
    Rosato, Antonello
    Succetti, Federico
    Barbirotta, Marcello
    Panella, Massimo
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,