Bilateral Memory Consolidation for Continual Learning

被引:2
|
作者
Nie, Xing [1 ,2 ]
Xu, Shixiong [1 ,2 ]
Liu, Xiyan [3 ]
Meng, Gaofeng [1 ,2 ,4 ]
Huo, Chunlei [1 ,2 ]
Xiang, Shiming [1 ,2 ]
机构
[1] Chinese Acad Sci, State Key Lab Multimodal Artificial Intelligence, Inst Automat, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing, Peoples R China
[3] Baidu Inc, Beijing, Peoples R China
[4] Chinese Acad Sci, HK Inst Sci & Innovat, Ctr Artificial Intelligence & Robot, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52729.2023.01538
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Humans are proficient at continuously acquiring and integrating new knowledge. By contrast, deep models forget catastrophically, especially when tackling highly long task sequences. Inspired by the way our brains constantly rewrite and consolidate past recollections, we propose a novel Bilateral Memory Consolidation (BiMeCo) framework that focuses on enhancing memory interaction capabilities. Specifically, BiMeCo explicitly decouples model parameters into short-term memory module and long-term memory module, responsible for representation ability of the model and generalization over all learned tasks, respectively. BiMeCo encourages dynamic interactions between two memory modules by knowledge distillation and momentum-based updating for forming generic knowledge to prevent forgetting. The proposed BiMeCo is parameter-efficient and can be integrated into existing methods seamlessly. Extensive experiments on challenging benchmarks show that BiMeCo significantly improves the performance of existing continual learning methods. For example, combined with the state-of-the-art method CwD [55], BiMeCo brings in significant gains of around 2% to 6% while using 2x fewer parameters on CIFAR-100 under ResNet-18.
引用
收藏
页码:16026 / 16035
页数:10
相关论文
共 50 条
  • [1] Dynamic Consolidation for Continual Learning
    Li, Hang
    Ma, Chen
    Chen, Xi
    Liu, Xue
    [J]. NEURAL COMPUTATION, 2023, 35 (02) : 228 - 248
  • [2] Policy Consolidation for Continual Reinforcement Learning
    Kaplanis, Christos
    Shanahan, Murray
    Clopath, Claudia
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [3] Meta-Consolidation for Continual Learning
    Joseph, K. J.
    Balasubramanian, Vineeth N.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [4] Continual Learning for Object Classification: Consolidation and Reconsolidation
    Turner, Daniel
    Cardoso, Pedro J. S.
    Rodrigues, Joao M. F.
    [J]. PERCEPTION, 2021, 50 (1_SUPPL) : 232 - 232
  • [5] Memory Bounds for Continual Learning
    Chen, Xi
    Papadimitriou, Christos
    Peng, Binghui
    [J]. 2022 IEEE 63RD ANNUAL SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE (FOCS), 2022, : 519 - 530
  • [6] Prediction Error-Driven Memory Consolidation for Continual Learning: On the Case of Adaptive Greenhouse Models
    Schillaci, Guido
    Schmidt, Uwe
    Miranda, Luis
    [J]. KUNSTLICHE INTELLIGENZ, 2021, 35 (01): : 71 - 80
  • [7] Prediction Error-Driven Memory Consolidation for Continual Learning: On the Case of Adaptive Greenhouse Models
    Guido Schillaci
    Uwe Schmidt
    Luis Miranda
    [J]. KI - Künstliche Intelligenz, 2021, 35 : 71 - 80
  • [8] Self-Paced Weight Consolidation for Continual Learning
    Cong, Wei
    Cong, Yang
    Sun, Gan
    Liu, Yuyang
    Dong, Jiahua
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (04) : 2209 - 2222
  • [9] Neural inhibition for continual learning and memory
    Barron, Helen C.
    [J]. CURRENT OPINION IN NEUROBIOLOGY, 2021, 67 : 85 - 94
  • [10] Online continual learning with declarative memory
    Xiao, Zhe
    Du, Zhekai
    Wang, Ruijin
    Gan, Ruimeng
    Li, Jingjing
    [J]. NEURAL NETWORKS, 2023, 163 : 146 - 155