Overcoming Catastrophic Forgetting Using Sparse Coding and Meta Learning

被引:8
|
作者
Hurtado, Julio [1 ]
Lobel, Hans [1 ,2 ]
Soto, Alvaro [1 ]
机构
[1] Pontificia Univ Catolica Chile, Dept Comp Sci, Santiago 7820436, Chile
[2] Pontificia Univ Catolica Chile, Dept Transport Engn & Logist, Santiago 782436, Chile
关键词
Task analysis; Interference; Training; Knowledge transfer; Adaptation models; Data models; Context modeling; Artificial intelligence; learning (artificial intelligence); machine learning; supervised learning; continual learning;
D O I
10.1109/ACCESS.2021.3090672
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Continuous learning occurs naturally in human beings. However, Deep Learning methods suffer from a problem known as Catastrophic Forgetting (CF) that consists of a model drastically decreasing its performance on previously learned tasks when it is sequentially trained on new tasks. This situation, known as task interference, occurs when a network modifies relevant weight values as it learns a new task. In this work, we propose two main strategies to face the problem of task interference in convolutional neural networks. First, we use a sparse coding technique to adaptively allocate model capacity to different tasks avoiding interference between them. Specifically, we use a strategy based on group sparse regularization to specialize groups of parameters to learn each task. Afterward, by adding binary masks, we can freeze these groups of parameters, using the rest of the network to learn new tasks. Second, we use a meta learning technique to foster knowledge transfer among tasks, encouraging weight reusability instead of overwriting. Specifically, we use an optimization strategy based on episodic training to foster learning weights that are expected to be useful to solve future tasks. Together, these two strategies help us to avoid interference by preserving compatibility with previous and future weight values. Using this approach, we achieve state-of-the-art results on popular benchmarks used to test techniques to avoid CF. In particular, we conduct an ablation study to identify the contribution of each component of the proposed method, demonstrating its ability to avoid retroactive interference with previous tasks and to promote knowledge transfer to future tasks.
引用
收藏
页码:88279 / 88290
页数:12
相关论文
共 50 条
  • [1] Complementary Learning for Overcoming Catastrophic Forgetting Using Experience Replay
    Rostami, Mohammad
    Kolouri, Soheil
    Pilly, Praveen K.
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 3339 - 3345
  • [2] Quantum Continual Learning Overcoming Catastrophic Forgetting
    Jiang, Wenjie
    Lu, Zhide
    Deng, Dong-Ling
    [J]. CHINESE PHYSICS LETTERS, 2022, 39 (05)
  • [3] Quantum Continual Learning Overcoming Catastrophic Forgetting
    蒋文杰
    鲁智徳
    邓东灵
    [J]. Chinese Physics Letters, 2022, 39 (05) : 29 - 41
  • [4] Knowledge Lock: Overcoming Catastrophic Forgetting in Federated Learning
    Wei, Guoyizhe
    Li, Xiu
    [J]. ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2022, PT I, 2022, 13280 : 601 - 612
  • [5] Overcoming Catastrophic Forgetting for Semantic Segmentation Via Incremental Learning
    Yang, Yizhuo
    Yuan, Shenghai
    Xie, Lihua
    [J]. 2022 17TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV), 2022, : 299 - 304
  • [6] PNSP: Overcoming catastrophic forgetting using Primary Null Space Projection in continual learning
    Zhou, DaiLiang
    Song, YongHong
    [J]. PATTERN RECOGNITION LETTERS, 2024, 179 : 137 - 143
  • [7] Overcoming catastrophic forgetting with classifier expander
    Liu, Xinchen
    Wang, Hongbo
    Tian, Yingjian
    Xie, Linyao
    [J]. ASIAN CONFERENCE ON MACHINE LEARNING, VOL 222, 2023, 222
  • [8] Overcoming catastrophic forgetting in neural networks
    Kirkpatricka, James
    Pascanu, Razvan
    Rabinowitz, Neil
    Veness, Joel
    Desjardins, Guillaume
    Rusu, Andrei A.
    Milan, Kieran
    Quan, John
    Ramalho, Tiago
    Grabska-Barwinska, Agnieszka
    Hassabis, Demis
    Clopath, Claudia
    Kumaran, Dharshan
    Hadsell, Raia
    [J]. PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2017, 114 (13) : 3521 - 3526
  • [9] Overcoming Catastrophic Forgetting in Continual Learning by Exploring Eigenvalues of Hessian Matrix
    Kong, Yajing
    Liu, Liu
    Chen, Huanhuan
    Kacprzyk, Janusz
    Tao, Dacheng
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, : 1 - 15
  • [10] Learn to Grow: A Continual Structure Learning Framework for Overcoming Catastrophic Forgetting
    Li, Xilai
    Zhou, Yingbo
    Wu, Tianfu
    Socher, Richard
    Xiong, Caiming
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97