Online Continual Learning via the Meta-learning update with Multi-scale Knowledge Distillation and Data Augmentation

被引:3
|
作者
Han, Ya-nan [1 ]
Liu, Jian-wei [1 ]
机构
[1] China Univ Petr, Coll Informat Sci & Engn, Dept Automat, Beijing, Peoples R China
关键词
Continual learning; The stability-plasticity dilemma; Meta-learning; Knowledge distillation; Data augmentation; NEURAL-NETWORKS;
D O I
10.1016/j.engappai.2022.104966
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Continual learning aims to rapidly and continually learn the current task from a sequence of tasks, using the knowledge obtained in the past, while performing well on prior tasks. A key challenge in this setting is the stability-plasticity dilemma existing in current and previous tasks, i.e., a high-stability network is weak to learn new knowledge in an effort to maintain previous knowledge. Correspondingly, a high-plasticity network can easily forget old tasks while dealing with well on the new task. Compared to other kinds of methods, the methods based on experience replay have shown great advantages to overcome catastrophic forgetting. One common limitation of this method is the data imbalance between the previous and current tasks, which would further aggravate forgetting. Moreover, how to effectively address the stability-plasticity dilemma in this setting is also an urgent problem to be solved. In this paper, we overcome these challenges by proposing a novel framework called Meta-learning update via Multi-scale Knowledge Distillation and Data Augmentation (MMKDDA). Specifically, we apply multi-scale knowledge distillation to grasp the evolution of long-range and short-range spatial relationships at different feature levels to alleviate the problem of data imbalance. Besides, our method mixes the samples from the episodic memory and current task in the online continual training procedure, thus alleviating the side influence due to the change of probability distribution. Moreover, we optimize our model via the meta-learning update by resorting to the number of tasks seen previously, which is helpful to keep a better balance between stability and plasticity. Finally, our extensive experiments on four benchmark datasets show the effectiveness of the proposed MMKDDA framework against other popular baselines, and ablation studies are also conducted to further analyze the role of each component in our framework.
引用
收藏
页数:17
相关论文
共 50 条
  • [41] Generalizable Speech Spoofing Detection Against Silence Trimming With Data Augmentation and Multi-Task Meta-Learning
    Wang, Li
    Yu, Lingyun
    Zhang, Yongdong
    Xie, Hongtao
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 3296 - 3310
  • [42] Online Knowledge Distillation via Mutual Contrastive Learning for Visual Recognition
    Yang, Chuanguang
    An, Zhulin
    Zhou, Helong
    Zhuang, Fuzhen
    Xu, Yongjun
    Zhang, Qian
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (08) : 10212 - 10227
  • [43] Data-Efficient Brain Connectome Analysis via Multi-Task Meta-Learning
    Yang, Yi
    Zhu, Yanqiao
    Cui, Hejie
    Kan, Xuan
    He, Lifang
    Guo, Ying
    Yang, Carl
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 4743 - 4751
  • [44] Dynamic multi-scale feature augmentation for inductive network representation learning
    Cui, Shicheng
    Li, Deqiang
    Zhang, Jing
    Pattern Recognition, 2025, 161
  • [45] Understanding Mobility via Deep Multi-Scale Learning
    Zhang, Rui
    Xie, Peng
    Wang, Chen
    Liu, Gaoyang
    Wan, Shaohua
    2018 INTERNATIONAL CONFERENCE ON IDENTIFICATION, INFORMATION AND KNOWLEDGE IN THE INTERNET OF THINGS, 2019, 147 : 487 - 494
  • [46] Handling Multi-scale Data via Multi-target Learning for Wind Speed Forecasting
    Appice, Annalisa
    Lanza, Antonietta
    Malerba, Donato
    FOUNDATIONS OF INTELLIGENT SYSTEMS (ISMIS 2018), 2018, 11177 : 357 - 366
  • [47] Learning to Select External Knowledge With Multi-Scale Negative Sampling
    He, Huang
    Lu, Hua
    Bao, Siqi
    Wang, Fan
    Wu, Hua
    Niu, Zheng-Yu
    Wang, Haifeng
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2024, 32 : 714 - 720
  • [48] Representation Learning of Knowledge Graphs with Multi-scale Capsule Network
    Cheng, Jingwei
    Yang, Zhi
    Dang, Jinming
    Pan, Chunguang
    Zhang, Fu
    INTELLIGENT DATA ENGINEERING AND AUTOMATED LEARNING - IDEAL 2019, PT I, 2019, 11871 : 282 - 290
  • [49] Online meta-learning approach for sensor fault diagnosis using limited data
    Wang, Lei
    Huang, Dukang
    Huang, Ke
    Civera, Marco
    SMART MATERIALS AND STRUCTURES, 2024, 33 (08)
  • [50] A multi-graph neural group recommendation model with meta-learning and multi-teacher distillation
    Zhou, Weizhen
    Huang, Zhenhua
    Wang, Cheng
    Chen, Yunwen
    KNOWLEDGE-BASED SYSTEMS, 2023, 276