Supervised Masked Knowledge Distillation for Few-Shot Transformers

被引:19
|
作者
Lin, Han [1 ]
Han, Guangxing [1 ]
Ma, Jiawei [1 ]
Huang, Shiyuan [1 ]
Lin, Xudong [1 ]
Chang, Shih-Fu [1 ]
机构
[1] Columbia Univ, New York, NY 10027 USA
关键词
D O I
10.1109/CVPR52729.2023.01882
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Vision Transformers (ViTs) emerge to achieve impressive performance on many data-abundant computer vision tasks by capturing long-range dependencies among local features. However, under few-shot learning (FSL) settings on small datasets with only a few labeled data, ViT tends to overfit and suffers from severe performance degradation due to its absence of CNN-alike inductive bias. Previous works in FSL avoid such problem either through the help of self-supervised auxiliary losses, or through the dextile uses of label information under supervised settings. But the gap between self-supervised and supervised few-shot Transformers is still unfilled. Inspired by recent advances in self-supervised knowledge distillation and masked image modeling (MIM), we propose a novel Supervised Masked Knowledge Distillation model (SMKD) for few-shot Transformers which incorporates label information into self-distillation frameworks. Compared with previous self-supervised methods, we allow intra-class knowledge distillation on both class and patch tokens, and introduce the challenging task of masked patch tokens reconstruction across intra-class images. Experimental results on four few-shot classification benchmark datasets show that our method with simple design outperforms previous methods by a large margin and achieves a new start-of-the-art. Detailed ablation studies confirm the effectiveness of each component of our model. Code for this paper is available here: https://github.com/HL-hanlin/SMKD.
引用
收藏
页码:19649 / 19659
页数:11
相关论文
共 50 条
  • [31] Prototype-wise self-knowledge distillation for few-shot segmentation
    Chen, Yadang
    Xu, Xinyu
    Wei, Chenchen
    Lu, Chuhan
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2024, 129
  • [32] Hyperbolic Insights With Knowledge Distillation for Cross-Domain Few-Shot Learning
    Yang, Xi
    Kong, Dechen
    Wang, Nannan
    Gao, Xinbo
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2025, 34 : 1921 - 1933
  • [33] AMP: Adaptive Masked Proxies for Few-Shot Segmentation
    Siam, Mennatullah
    Oreshkin, Boris N.
    Jagersand, Martin
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 5248 - 5257
  • [34] Few-Shot Knowledge Graph Completion
    Zhang, Chuxu
    Yao, Huaxiu
    Huang, Chao
    Jiang, Meng
    Li, Zhenhui
    Chawla, Nitesh, V
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 3041 - 3048
  • [35] Few-Shot Graph Anomaly Detection via Dual-Level Knowledge Distillation
    Li, Xuan
    Cheng, Dejie
    Zhang, Luheng
    Zhang, Chengfang
    Feng, Ziliang
    ENTROPY, 2025, 27 (01)
  • [36] Few-shot image classification with improved similarity relationships in self-knowledge distillation
    Li, Liang
    Jin, Weidong
    Ren, Junxiao
    Huang, Yingkun
    Yan, Kang
    2022 41ST CHINESE CONTROL CONFERENCE (CCC), 2022, : 7053 - 7058
  • [37] Contrastive knowledge-augmented self-distillation approach for few-shot learning
    Zhang, Lixu
    Shao, Mingwen
    Chen, Sijie
    Liu, Fukang
    JOURNAL OF ELECTRONIC IMAGING, 2023, 32 (05)
  • [38] KDNet: Leveraging Vision-Language Knowledge Distillation for Few-Shot Object Detection
    Ma, Mengyuan
    Qian, Lin
    Yin, Hujun
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING-ICANN 2024, PT II, 2024, 15017 : 153 - 167
  • [39] Semantic-aware Knowledge Distillation for Few-Shot Class-Incremental Learning
    Cheraghian, Ali
    Rahman, Shafin
    Fang, Pengfei
    Roy, Soumava Kumar
    Petersson, Lars
    Harandi, Mehrtash
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 2534 - 2543
  • [40] Enhancing the Generalization Performance of Few-Shot Image Classification with Self-Knowledge Distillation
    Li, Liang
    Jin, Weidong
    Huang, Yingkun
    Ren, Junxiao
    STUDIES IN INFORMATICS AND CONTROL, 2022, 31 (02): : 71 - 80