Preempting Catastrophic Forgetting in Continual Learning Models by Anticipatory Regularization

被引:0
|
作者
El Khatib, Alaa [1 ]
Karray, Fakhri [1 ]
机构
[1] Univ Waterloo, Elect & Comp Engn, Waterloo, ON, Canada
关键词
D O I
10.1109/ijcnn.2019.8852426
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural networks trained on tasks sequentially tend to degrade in performance, on the average, the more tasks they see, as the representations learned for one task get progressively modified while learning subsequent tasks. This phenomenon-known as catastrophic forgetting-is a major obstacle on the road toward designing agents that can continually learn new concepts and tasks the way, say, humans do. A common approach to containing catastrophic forgetting is to use regularization to slow down learning on weights deemed important to previously learned tasks. We argue in this paper that, on their own, such post hoc measures to safeguard what has been learned can, even in their more sophisticated variants, paralyze the network and degrade its capacity to learn and counter forgetting as the number of tasks learned increases. We propose insteador possibly in conjunction-that, in anticipation of future tasks, regularization be applied to drive the optimization of network weights toward reusable solutions. We show that one way to achieve this is through an auxiliary unsupervised reconstruction loss that encourages the learned representations not only to be useful for solving, say, the current classification task, but also to reflect the content of the data being processed-content that is generally richer than it is discriminative for any one task. We compare our approach to the recent elastic weight consolidation regularization approach, and show that, although we do not explicitly try to preserve important weights or pass on any information about the data distribution of learned tasks, our model is comparable in performance, and in some cases better.
引用
收藏
页数:7
相关论文
共 50 条
  • [31] Overcoming Catastrophic Forgetting in Continual Fine-Grained Urban Flow Inference
    Xu, Xovee
    Zhong, Ting
    Yu, Haoyang
    Zhou, Fan
    Trajcevski, Goce
    ACM TRANSACTIONS ON SPATIAL ALGORITHMS AND SYSTEMS, 2024, 10 (04)
  • [32] A Comprehensive Survey of Forgetting in Deep Learning Beyond Continual Learning
    Wang, Zhenyi
    Yang, Enneng
    Shen, Li
    Huang, Heng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2025, 47 (03) : 1464 - 1483
  • [33] Catastrophic Forgetting in Deep Learning: A Comprehensive Taxonomy
    Aleixo, Everton Lima
    Colonna, Juan G.
    Cristo, Marco
    Fernandes, Everlandio
    Journal of the Brazilian Computer Society, 2024, 30 (01) : 175 - 211
  • [34] Mitigating Catastrophic Forgetting in Robot Continual Learning: A Guided Policy Search Approach Enhanced With Memory-Aware Synapses
    Dong, Qingwei
    Zeng, Peng
    He, Yunpeng
    Wan, Guangxi
    Dong, Xiaoting
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (12): : 11242 - 11249
  • [35] Mitigating Catastrophic Forgetting with Complementary Layered Learning
    Mondesire, Sean
    Wiegand, R. Paul
    ELECTRONICS, 2023, 12 (03)
  • [36] Comparative Analysis of Catastrophic Forgetting in Metric Learning
    Huo, Jiahao
    van Zyl, Terence L.
    2020 7TH INTERNATIONAL CONFERENCE ON SOFT COMPUTING & MACHINE INTELLIGENCE (ISCMI 2020), 2020, : 68 - 72
  • [37] Reducing catastrophic forgetting with learning on synthetic data
    Masarczyk, Wojciech
    Tautkute, Ivona
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 1019 - 1024
  • [38] On Reinforcement Learning and Distribution Matching for Fine-Tuning Language Models with no Catastrophic Forgetting
    Korbak, Tomasz
    Elsahar, Hady
    Kruszewski, German
    Dymetman, Marc
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [39] InfoCL: Alleviating Catastrophic Forgetting in Continual Text Classification from An Information Theoretic Perspective
    Song, Yifan
    Wang, Peiyi
    Xiong, Weimin
    Zhu, Dawei
    Liu, Tianyu
    Sui, Zhifang
    Li, Sujian
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 14557 - 14570
  • [40] PLOP: Learning without Forgetting for Continual Semantic Segmentation
    Douillard, Arthur
    Chen, Yifu
    Dapogny, Arnaud
    Cord, Matthieu
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 4039 - 4049