Class similarity weighted knowledge distillation for few shot incremental learning

被引:2
|
作者
Akmel, Feidu [1 ]
Meng, Fanman [1 ]
Wu, Qingbo [1 ]
Chen, Shuai [1 ]
Zhang, Runtong [1 ]
Assefa, Maregu [2 ]
机构
[1] Univ Elect Sci & Technol China, Sch Informat & Commun Engn, Chengdu, Peoples R China
[2] Univ Elect Sci & Technol China, Sch Informat & Software Engn, Chengdu, Peoples R China
关键词
Knowledge distillation; Semantic information; Few shot; Incremental learning;
D O I
10.1016/j.neucom.2024.127587
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Few -shot class incremental learning illustrates the challenges of learning new concepts, where the learner can access only a small sample per concept. The standard incremental learning techniques cannot be applied directly because of the small number of samples for training. Moreover, catastrophic forgetting is the propensity of an Artificial Neural Network to fully and abruptly forget previously learned knowledge upon learning new knowledge. This problem happens due to a lack of supervision in older classes or an imbalance between the old and new classes. In this work, we propose a new distillation structure to tackle the forgetting and overfitting issues. Particularly, we suggest a dual distillation module that adaptably draws knowledge from two different but complementary teachers. The first teacher is the base model, which has been trained on large class data, and the second teacher is the updated model from the previous K-1 session, which contains the modified knowledge of previously observed new classes. Thus, the first teacher can reduce overfitting issues by transferring the knowledge obtained from the base classes to the new classes. While the second teacher can reduce knowledge forgetting by distilling knowledge from the previous model. Additionally, we use semantic information as word embedding to facilitate the distillation process. To align visual and semantic vectors, we used the attention mechanism of the embedding of visual data. With extensive experiments on different data sets such as Mini-ImageNet, CIFAR100, and CUB200, our model shows state-of-the-art performance compared to the existing few shot incremental learning methods.
引用
收藏
页数:11
相关论文
共 50 条
  • [41] M2SD:Multiple Mixing Self-Distillation for Few-Shot Class-Incremental Learning
    Lin, Jinhao
    Wu, Ziheng
    Lin, Weifeng
    Huang, Jun
    Luo, RongHua
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 4, 2024, : 3422 - 3431
  • [42] Learnable Distribution Calibration for Few-Shot Class-Incremental Learning
    Liu, Binghao
    Yang, Boyu
    Xie, Lingxi
    Wang, Ren
    Tian, Qi
    Ye, Qixiang
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (10) : 12699 - 12706
  • [43] Model Attention Expansion for Few-Shot Class-Incremental Learning
    Wang, Xuan
    Ji, Zhong
    Yu, Yunlong
    Pang, Yanwei
    Han, Jungong
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 4419 - 4431
  • [44] A Few-Shot Class-Incremental Learning Approach for Intrusion Detection
    Wang, Tingting
    Lv, Qiujian
    Hu, Bo
    Sun, Degang
    30TH INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATIONS AND NETWORKS (ICCCN 2021), 2021,
  • [45] SEMI-SUPERVISED FEW-SHOT CLASS-INCREMENTAL LEARNING
    Cui, Yawen
    Xiong, Wuti
    Tavakolian, Mohammad
    Liu, Li
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 1239 - 1243
  • [46] Few Shot Class Incremental Learning via Efficient Prototype Replay and Calibration
    Zhang, Wei
    Gu, Xiaodong
    ENTROPY, 2023, 25 (05)
  • [47] Multimodal Parameter-Efficient Few-Shot Class Incremental Learning
    D'Alessandro, Marco
    Alonso, Alberto
    Calabres, Enrique
    Galar, Mikel
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 3385 - 3395
  • [48] Memorizing Complementation Network for Few-Shot Class-Incremental Learning
    Ji, Zhong
    Hou, Zhishen
    Liu, Xiyao
    Pang, Yanwei
    Li, Xuelong
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 937 - 948
  • [49] Efficient-PrototypicalNet with self knowledge distillation for few-shot learning
    Lim, Jit Yan
    Lim, Kian Ming
    Ooi, Shih Yin
    Lee, Chin Poo
    NEUROCOMPUTING, 2021, 459 : 327 - 337
  • [50] Filter Bank Networks for Few-Shot Class-Incremental Learning
    Zhou, Yanzhao
    Liu, Binghao
    Liu, Yiran
    Jiao, Jianbin
    CMES-COMPUTER MODELING IN ENGINEERING & SCIENCES, 2023, 137 (01): : 647 - 668