Overcoming Catastrophic Forgetting via Direction-Constrained Optimization

被引:0
|
作者
Teng, Yunfei [1 ]
Choromanska, Anna [1 ]
Campbell, Murray [2 ]
Lu, Songtao [2 ]
Ram, Parikshit [2 ]
Horesh, Lior [2 ]
机构
[1] NYU, New York, NY 10012 USA
[2] IBM Res, New York, NY USA
关键词
Continual/Lifelong learning; Deep learning; Optimization;
D O I
10.1007/978-3-031-26387-3_41
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper studies a new design of the optimization algorithm for training deep learning models with a fixed architecture of the classification network in a continual learning framework. The training data is non-stationary and the non-stationarity is imposed by a sequence of distinct tasks. We first analyze a deep model trained on only one learning task in isolation and identify a region in network parameter space, where the model performance is close to the recovered optimum. We provide empirical evidence that this region resembles a cone that expands along the convergence direction. We study the principal directions of the trajectory of the optimizer after convergence and show that traveling along a few top principal directions can quickly bring the parameters outside the cone but this is not the case for the remaining directions. We argue that catastrophic forgetting in a continual learning setting can be alleviated when the parameters are constrained to stay within the intersection of the plausible cones of individual tasks that were so far encountered during training. Based on this observation we present our direction-constrained optimization (DCO) method, where for each task we introduce a linear autoencoder to approximate its corresponding top forbidden principal directions. They are then incorporated into the loss function in the form of a regularization term for the purpose of learning the coming tasks without forgetting. Furthermore, in order to control the memory growth as the number of tasks increases, we propose a memory-efficient version of our algorithm called compressed DCO (DCO-COMP) that allocates a memory of fixed size for storing all autoencoders. We empirically demonstrate that our algorithm performs favorably compared to other state-of-art regularization-based continual learning methods. The codes are publicly available at https://github.com/yunfei-teng/DCO.
引用
收藏
页码:675 / 692
页数:18
相关论文
共 50 条
  • [1] Direction-Constrained Rectangle Escape Routing
    Yan, Jin-Tai
    [J]. ACM TRANSACTIONS ON DESIGN AUTOMATION OF ELECTRONIC SYSTEMS, 2018, 23 (03)
  • [2] Overcoming Catastrophic Forgetting for Semantic Segmentation Via Incremental Learning
    Yang, Yizhuo
    Yuan, Shenghai
    Xie, Lihua
    [J]. 2022 17TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV), 2022, : 299 - 304
  • [3] Overcoming catastrophic forgetting with classifier expander
    Liu, Xinchen
    Wang, Hongbo
    Tian, Yingjian
    Xie, Linyao
    [J]. ASIAN CONFERENCE ON MACHINE LEARNING, VOL 222, 2023, 222
  • [4] Overcoming catastrophic forgetting in neural networks
    Kirkpatricka, James
    Pascanu, Razvan
    Rabinowitz, Neil
    Veness, Joel
    Desjardins, Guillaume
    Rusu, Andrei A.
    Milan, Kieran
    Quan, John
    Ramalho, Tiago
    Grabska-Barwinska, Agnieszka
    Hassabis, Demis
    Clopath, Claudia
    Kumaran, Dharshan
    Hadsell, Raia
    [J]. PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2017, 114 (13) : 3521 - 3526
  • [5] Overcoming Catastrophic Forgetting in Incremental Object Detection via Elastic Response Distillation
    Feng, Tao
    Wang, Mang
    Yuan, Hangjie
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 9417 - 9426
  • [6] Overcoming Catastrophic Forgetting by Bayesian Generative Regularization
    Chen, Patrick H.
    Wei, Wei
    Hsieh, Cho-jui
    Dai, Bo
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [7] Quantum Continual Learning Overcoming Catastrophic Forgetting
    Jiang, Wenjie
    Lu, Zhide
    Deng, Dong-Ling
    [J]. CHINESE PHYSICS LETTERS, 2022, 39 (05)
  • [8] Quantum Continual Learning Overcoming Catastrophic Forgetting
    蒋文杰
    鲁智徳
    邓东灵
    [J]. Chinese Physics Letters, 2022, 39 (05) : 29 - 41
  • [9] Overcoming Catastrophic Forgetting in Graph Neural Networks
    Liu, Huihui
    Yang, Yiding
    Wang, Xinchao
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8653 - 8661
  • [10] Overcoming Catastrophic Forgetting with Gaussian Mixture Replay
    Pfuelb, Benedikt
    Gepperth, Alexander
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,