Multi-granularity knowledge distillation and prototype consistency regularization for class-incremental learning

被引:9
|
作者
Shi, Yanyan [1 ]
Shi, Dianxi [2 ]
Qiao, Ziteng [2 ]
Wang, Zhen [2 ]
Zhang, Yi [2 ]
Yang, Shaowu [1 ]
Qiu, Chunping [2 ]
机构
[1] Natl Univ Def Technol, Coll Comp, Changsha 410073, Peoples R China
[2] Natl Innovat Inst Def Technol, Beijing 100071, Peoples R China
基金
中国国家自然科学基金;
关键词
Class -incremental learning; Knowledge distillation; Consistency regularization; Image classification;
D O I
10.1016/j.neunet.2023.05.006
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks (DNNs) are prone to the notorious catastrophic forgetting problem when learning new tasks incrementally. Class-incremental learning (CIL) is a promising solution to tackle the challenge and learn new classes while not forgetting old ones. Existing CIL approaches adopted stored representative exemplars or complex generative models to achieve good performance. However, storing data from previous tasks causes memory or privacy issues, and the training of generative models is unstable and inefficient. This paper proposes a method based on multi-granularity knowledge distillation and prototype consistency regularization (MDPCR) that performs well even when the previous training data is unavailable. First, we propose to design knowledge distillation losses in the deep feature space to constrain the incremental model trained on the new data. Thereby, multi -granularity is captured from three aspects: by distilling multi-scale self-attentive features, the feature similarity probability, and global features to maximize the retention of previous knowledge, effectively alleviating catastrophic forgetting. Conversely, we preserve the prototype of each old class and employ prototype consistency regularization (PCR) to ensure that the old prototypes and semantically enhanced prototypes produce consistent prediction, which excels in enhancing the robustness of old prototypes and reduces the classification bias. Extensive experiments on three CIL benchmark datasets confirm that MDPCR performs significantly better over exemplar-free methods and outperforms typical exemplar-based approaches.(c) 2023 Elsevier Ltd. All rights reserved.
引用
收藏
页码:617 / 630
页数:14
相关论文
共 50 条
  • [21] Flexible few-shot class-incremental learning with prototype container
    Xu, Xinlei
    Wang, Zhe
    Fu, Zhiling
    Guo, Wei
    Chi, Ziqiu
    Li, Dongdong
    [J]. NEURAL COMPUTING & APPLICATIONS, 2023, 35 (15): : 10875 - 10889
  • [22] CLASS-INCREMENTAL LEARNING WITH REPETITION
    Hemati, Hamed
    Cossu, Andrea
    Carta, Antonio
    Hurtado, Julio
    Pellegrini, Lorenzo
    Bacciu, Davide
    Lomonaco, Vincenzo
    Borth, Damian
    [J]. CONFERENCE ON LIFELONG LEARNING AGENTS, VOL 232, 2023, 232 : 437 - 455
  • [23] AN END-TO-END ARCHITECTURE FOR CLASS-INCREMENTAL OBJECT DETECTION WITH KNOWLEDGE DISTILLATION
    Hao, Yu
    Fu, Yanwei
    Jiang, Yu-Gang
    Tian, Qi
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2019, : 1 - 6
  • [24] Class-Incremental Learning: A Survey
    Zhou, Da-Wei
    Wang, Qi-Wei
    Qi, Zhi-Hong
    Ye, Han-Jia
    Zhan, De-Chuan
    Liu, Ziwei
    [J]. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2024, 46 (12) : 9851 - 9873
  • [25] Online Multi-Granularity Distillation for GAN Compression
    Ren, Yuxi
    Wu, Jie
    Xiao, Xuefeng
    Yang, Jianchao
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 6773 - 6783
  • [26] Federated Class-Incremental Learning
    Dong, Jiahua
    Wang, Lixu
    Fang, Zhen
    Sun, Gan
    Xu, Shichao
    Wang, Xiao
    Zhu, Qi
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 10154 - 10163
  • [27] Dataset Knowledge Transfer for Class-Incremental Learning without Memory
    Slim, Habib
    Belouadah, Eden
    Popescu, Adrian
    Onchis, Darian
    [J]. 2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), 2022, : 3311 - 3320
  • [28] Knowledge Transfer between Multi-granularity Models for Reinforcement Learning
    Wang, Lan
    Tang, Kaiqiang
    Xin, Bo
    Chen, Chunlin
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2018, : 2881 - 2886
  • [29] MULTI-GRANULARITY KNOWLEDGE MINING ON THE WEB
    Xie, Ming
    [J]. INTERNATIONAL JOURNAL OF SOFTWARE ENGINEERING AND KNOWLEDGE ENGINEERING, 2012, 22 (01) : 1 - 16
  • [30] Uncertainty-Guided Semi-Supervised Few-Shot Class-Incremental Learning With Knowledge Distillation
    Cui, Yawen
    Deng, Wanxia
    Xu, Xin
    Liu, Zhen
    Liu, Zhong
    Pietikainen, Matti
    Liu, Li
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 6422 - 6435