Supervised Masked Knowledge Distillation for Few-Shot Transformers

被引:19
|
作者
Lin, Han [1 ]
Han, Guangxing [1 ]
Ma, Jiawei [1 ]
Huang, Shiyuan [1 ]
Lin, Xudong [1 ]
Chang, Shih-Fu [1 ]
机构
[1] Columbia Univ, New York, NY 10027 USA
关键词
D O I
10.1109/CVPR52729.2023.01882
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Vision Transformers (ViTs) emerge to achieve impressive performance on many data-abundant computer vision tasks by capturing long-range dependencies among local features. However, under few-shot learning (FSL) settings on small datasets with only a few labeled data, ViT tends to overfit and suffers from severe performance degradation due to its absence of CNN-alike inductive bias. Previous works in FSL avoid such problem either through the help of self-supervised auxiliary losses, or through the dextile uses of label information under supervised settings. But the gap between self-supervised and supervised few-shot Transformers is still unfilled. Inspired by recent advances in self-supervised knowledge distillation and masked image modeling (MIM), we propose a novel Supervised Masked Knowledge Distillation model (SMKD) for few-shot Transformers which incorporates label information into self-distillation frameworks. Compared with previous self-supervised methods, we allow intra-class knowledge distillation on both class and patch tokens, and introduce the challenging task of masked patch tokens reconstruction across intra-class images. Experimental results on four few-shot classification benchmark datasets show that our method with simple design outperforms previous methods by a large margin and achieves a new start-of-the-art. Detailed ablation studies confirm the effectiveness of each component of our model. Code for this paper is available here: https://github.com/HL-hanlin/SMKD.
引用
收藏
页码:19649 / 19659
页数:11
相关论文
共 50 条
  • [21] Few-shot class incremental learning via prompt transfer and knowledge distillation
    Akmel, Feidu
    Meng, Fanman
    Liu, Mingyu
    Zhang, Runtong
    Teka, Asebe
    Lemuye, Elias
    IMAGE AND VISION COMPUTING, 2024, 151
  • [22] Self-Training Based Few-Shot Node Classification by Knowledge Distillation
    Wu, Zongqian
    Mo, Yujie
    Zhou, Peng
    Yuan, Shangbo
    Zhu, Xiaofeng
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 14, 2024, : 15988 - 15995
  • [23] Global and Local Knowledge Distillation Method for Few-Shot Classification of Electrical Equipment
    Zhou, Bojun
    Zhao, Jiahao
    Yan, Chunkai
    Zhang, Xinsong
    Gu, Juping
    APPLIED SCIENCES-BASEL, 2023, 13 (12):
  • [24] Joint data augmentation and knowledge distillation for few-shot continual relation extraction
    Wei, Zhongcheng
    Zhang, Yunping
    Lian, Bin
    Fan, Yongjian
    Zhao, Jijun
    APPLIED INTELLIGENCE, 2024, 54 (04) : 3516 - 3528
  • [25] Progressive Network Grafting With Local Features Embedding for Few-Shot Knowledge Distillation
    Du, Weidong
    IEEE ACCESS, 2022, 10 : 116196 - 116204
  • [26] Overcoming Granularity Mismatch in Knowledge Distillation for Few-Shot Hyperspectral Image Classification
    Wu, Hao
    Xue, Zhaohui
    Zhou, Shaoguang
    Su, Hongjun
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2025, 63
  • [27] Few-Shot Class-Incremental Learning via Relation Knowledge Distillation
    Dong, Songlin
    Hong, Xiaopeng
    Tao, Xiaoyu
    Chang, Xinyuan
    Wei, Xing
    Gong, Yihong
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 1255 - 1263
  • [28] Joint data augmentation and knowledge distillation for few-shot continual relation extraction
    Zhongcheng Wei
    Yunping Zhang
    Bin Lian
    Yongjian Fan
    Jijun Zhao
    Applied Intelligence, 2024, 54 : 3516 - 3528
  • [29] Fully Self-Supervised Out-of-Domain Few-Shot Learning with Masked Autoencoders
    Walsh, Reece
    Osman, Islam
    Abdelaziz, Omar
    Shehata, Mohamed S.
    JOURNAL OF IMAGING, 2024, 10 (01)
  • [30] Few-Shot Learning Meets Transformer: Unified Query-Support Transformers for Few-Shot Classification
    Wang, Xixi
    Wang, Xiao
    Jiang, Bo
    Luo, Bin
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (12) : 7789 - 7802