Online Continual Learning via the Meta-learning update with Multi-scale Knowledge Distillation and Data Augmentation

被引:3
|
作者
Han, Ya-nan [1 ]
Liu, Jian-wei [1 ]
机构
[1] China Univ Petr, Coll Informat Sci & Engn, Dept Automat, Beijing, Peoples R China
关键词
Continual learning; The stability-plasticity dilemma; Meta-learning; Knowledge distillation; Data augmentation; NEURAL-NETWORKS;
D O I
10.1016/j.engappai.2022.104966
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Continual learning aims to rapidly and continually learn the current task from a sequence of tasks, using the knowledge obtained in the past, while performing well on prior tasks. A key challenge in this setting is the stability-plasticity dilemma existing in current and previous tasks, i.e., a high-stability network is weak to learn new knowledge in an effort to maintain previous knowledge. Correspondingly, a high-plasticity network can easily forget old tasks while dealing with well on the new task. Compared to other kinds of methods, the methods based on experience replay have shown great advantages to overcome catastrophic forgetting. One common limitation of this method is the data imbalance between the previous and current tasks, which would further aggravate forgetting. Moreover, how to effectively address the stability-plasticity dilemma in this setting is also an urgent problem to be solved. In this paper, we overcome these challenges by proposing a novel framework called Meta-learning update via Multi-scale Knowledge Distillation and Data Augmentation (MMKDDA). Specifically, we apply multi-scale knowledge distillation to grasp the evolution of long-range and short-range spatial relationships at different feature levels to alleviate the problem of data imbalance. Besides, our method mixes the samples from the episodic memory and current task in the online continual training procedure, thus alleviating the side influence due to the change of probability distribution. Moreover, we optimize our model via the meta-learning update by resorting to the number of tasks seen previously, which is helpful to keep a better balance between stability and plasticity. Finally, our extensive experiments on four benchmark datasets show the effectiveness of the proposed MMKDDA framework against other popular baselines, and ablation studies are also conducted to further analyze the role of each component in our framework.
引用
下载
收藏
页数:17
相关论文
共 50 条
  • [1] Data Augmentation for Meta-Learning
    Ni, Renkun
    Goldblum, Micah
    Sharaf, Amr
    Kong, Kezhi
    Goldstein, Tom
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [2] Reconciling meta-learning and continual learning with online mixtures of tasks
    Jerfel, Ghassen
    Grant, Erin
    Griffiths, Thomas L.
    Heller, Katherine
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [3] Adaptive Multi-Teacher Knowledge Distillation with Meta-Learning
    Zhang, Hailin
    Chen, Defang
    Wang, Can
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1943 - 1948
  • [4] A rolling bearing fault diagnosis method based on multi-scale knowledge distillation and continual learning
    Xia, Yifei
    Gao, Jun
    Shao, Xing
    Wang, Cuixiang
    Zhendong yu Chongji/Journal of Vibration and Shock, 2024, 43 (12): : 276 - 285
  • [5] Continual Quality Estimation with Online Bayesian Meta-Learning
    Obamuyide, Abiola
    Fomicheva, Marina
    Specia, Lucia
    ACL-IJCNLP 2021: THE 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 2, 2021, : 190 - 197
  • [6] Knowledge distillation via adaptive meta-learning for graph neural network
    Shen, Tiesunlong
    Wang, Jin
    Zhang, Xuejie
    INFORMATION SCIENCES, 2025, 689
  • [7] Meta-Learning for Online Update of Recommender Systems
    Kim, Minseok
    Song, Hwanjun
    Shin, Yooju
    Park, Dongmin
    Shin, Kijung
    Lee, Jae-Gil
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 4065 - 4074
  • [8] A Personalized Federated Learning Algorithm Based on Meta-Learning and Knowledge Distillation
    Sun Y.
    Shi Y.
    Wang Z.
    Li M.
    Si P.
    Beijing Youdian Daxue Xuebao/Journal of Beijing University of Posts and Telecommunications, 2023, 46 (01): : 12 - 18
  • [9] Knowledge Distillation for Model-Agnostic Meta-Learning
    Zhang, Min
    Wang, Donglin
    Gai, Sibo
    ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 1355 - 1362
  • [10] Multi-scale Relation Network for Few-Shot Learning Based on Meta-learning
    Ding, Yueming
    Tian, Xia
    Yin, Lirong
    Chen, Xiaobing
    Liu, Shan
    Yang, Bo
    Zheng, Wenfeng
    COMPUTER VISION SYSTEMS (ICVS 2019), 2019, 11754 : 343 - 352