Bilateral Memory Consolidation for Continual Learning

被引:2
|
作者
Nie, Xing [1 ,2 ]
Xu, Shixiong [1 ,2 ]
Liu, Xiyan [3 ]
Meng, Gaofeng [1 ,2 ,4 ]
Huo, Chunlei [1 ,2 ]
Xiang, Shiming [1 ,2 ]
机构
[1] Chinese Acad Sci, State Key Lab Multimodal Artificial Intelligence, Inst Automat, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Artificial Intelligence, Beijing, Peoples R China
[3] Baidu Inc, Beijing, Peoples R China
[4] Chinese Acad Sci, HK Inst Sci & Innovat, Ctr Artificial Intelligence & Robot, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR52729.2023.01538
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Humans are proficient at continuously acquiring and integrating new knowledge. By contrast, deep models forget catastrophically, especially when tackling highly long task sequences. Inspired by the way our brains constantly rewrite and consolidate past recollections, we propose a novel Bilateral Memory Consolidation (BiMeCo) framework that focuses on enhancing memory interaction capabilities. Specifically, BiMeCo explicitly decouples model parameters into short-term memory module and long-term memory module, responsible for representation ability of the model and generalization over all learned tasks, respectively. BiMeCo encourages dynamic interactions between two memory modules by knowledge distillation and momentum-based updating for forming generic knowledge to prevent forgetting. The proposed BiMeCo is parameter-efficient and can be integrated into existing methods seamlessly. Extensive experiments on challenging benchmarks show that BiMeCo significantly improves the performance of existing continual learning methods. For example, combined with the state-of-the-art method CwD [55], BiMeCo brings in significant gains of around 2% to 6% while using 2x fewer parameters on CIFAR-100 under ResNet-18.
引用
下载
收藏
页码:16026 / 16035
页数:10
相关论文
共 50 条
  • [21] Evolving Parameterized Prompt Memory for Continual Learning
    Kurniawan, Muhammad Rifki
    Song, Xiang
    Ma, Zhiheng
    He, Yuhang
    Gong, Yihong
    Yang, Qi
    Wei, Xing
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 12, 2024, : 13301 - 13309
  • [22] SYNERGY BETWEEN SYNAPTIC CONSOLIDATION AND EXPERIENCE REPLAY FOR GENERAL CONTINUAL LEARNING
    Sarfraz, Fahad
    Arani, Elahe
    Zonooz, Bahram
    CONFERENCE ON LIFELONG LEARNING AGENTS, VOL 199, 2022, 199
  • [23] Sparsified Subgraph Memory for Continual Graph Representation Learning
    Zhang, Xikun
    Song, Dongjin
    Tao, Dacheng
    2022 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2022, : 1335 - 1340
  • [24] Memory Population in Continual Learning via Outlier Elimination
    Hurtado, Julio
    Raymond-Saez, Alain
    Araujo, Vladimir
    Lomonaco, Vincenzo
    Soto, Alvaro
    Bacciu, Davide
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 3473 - 3482
  • [25] Continual learning via region-aware memory
    Zhao, Kai
    Fu, Zhenyong
    Yang, Jian
    APPLIED INTELLIGENCE, 2023, 53 (07) : 8389 - 8401
  • [26] Convolutional Neural Network With Developmental Memory for Continual Learning
    Park, Gyeong-Moon
    Yoo, Sahng-Min
    Kim, Jong-Hwan
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (06) : 2691 - 2705
  • [27] Federated probability memory recall for federated continual learning
    Wang, Zhe
    Zhang, Yu
    Xu, Xinlei
    Fu, Zhiling
    Yang, Hai
    Du, Wenli
    INFORMATION SCIENCES, 2023, 629 : 551 - 565
  • [28] Prototype-Guided Memory Replay for Continual Learning
    Ho, Stella
    Liu, Ming
    Du, Lan
    Gao, Longxiang
    Xiang, Yong
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (08) : 10973 - 10983
  • [29] Motor memory: Representation, learning and consolidation
    Zabkar, Jure
    Leonardis, Ales
    BIOLOGICALLY INSPIRED COGNITIVE ARCHITECTURES, 2016, 16 : 64 - 74
  • [30] Astrocytic involvement in learning and memory consolidation
    Gibbs, Marie E.
    Hutchinson, Dana
    Hertz, Leif
    NEUROSCIENCE AND BIOBEHAVIORAL REVIEWS, 2008, 32 (05): : 927 - 944