Rethinking Generalization in Few-Shot Classification

被引:0
|
作者
Hiller, Markus [1 ]
Ma, Rongkai [2 ]
Harandi, Mehrtash [2 ]
Drummond, Tom [1 ]
机构
[1] Univ Melbourne, Sch Comp & Informat Syst, Melbourne, Vic, Australia
[2] Monash Univ, Dept Elect & Comp Syst Engn, Clayton, Vic, Australia
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Single image-level annotations only correctly describe an often small subset of an image's content, particularly when complex real-world scenes are depicted. While this might be acceptable in many classification scenarios, it poses a significant challenge for applications where the set of classes differs significantly between training and test time. In this paper, we take a closer look at the implications in the context of few-shot learning. Splitting the input samples into patches and encoding these via the help of Vision Transformers allows us to establish semantic correspondences between local regions across images and independent of their respective class. The most informative patch embeddings for the task at hand are then determined as a function of the support set via online optimization at inference time, additionally providing visual interpretability of 'what matters most' in the image. We build on recent advances in unsupervised training of networks via masked image modelling to overcome the lack of fine-grained labels and learn the more general statistical structure of the data while avoiding negative image-level annotation influence, aka supervision collapse. Experimental results show the competitiveness of our approach, achieving new state-of-the-art results on four popular few-shot classification benchmarks for 5-shot and 1-shot scenarios.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] Few-Shot Learning Meets Transformer: Unified Query-Support Transformers for Few-Shot Classification
    Wang, Xixi
    Wang, Xiao
    Jiang, Bo
    Luo, Bin
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (12) : 7789 - 7802
  • [22] Multiscale attention for few-shot image classification
    Zhou, Tong
    Dong, Changyin
    Song, Junshu
    Zhang, Zhiqiang
    Wang, Zhen
    Chang, Bo
    Chen, Dechun
    [J]. COMPUTATIONAL INTELLIGENCE, 2024, 40 (02)
  • [23] Few-shot classification with Fork Attention Adapter
    Sun, Jieqi
    Li, Jian
    [J]. PATTERN RECOGNITION, 2024, 156
  • [24] Few-shot learning for short text classification
    Yan, Leiming
    Zheng, Yuhui
    Cao, Jie
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2018, 77 (22) : 29799 - 29810
  • [25] Adversarial training for few-shot text classification
    Croce, Danilo
    Castellucci, Giuseppe
    Basili, Roberto
    [J]. INTELLIGENZA ARTIFICIALE, 2020, 14 (02) : 201 - 214
  • [26] Few-Shot Classification with Semantic Augmented Activators
    Gao, Ruixuan
    Su, Han
    Tang, Peisen
    [J]. PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT VIII, 2024, 14432 : 340 - 352
  • [27] Decision fusion for few-shot image classification
    Tianhao Yuan
    Weifeng Liu
    Fei Yan
    Baodi Liu
    [J]. International Journal of Multimedia Information Retrieval, 2023, 12
  • [28] Cross Attention Network for Few-shot Classification
    Hou, Ruibing
    Chang, Hong
    Ma, Bingpeng
    Shan, Shiguang
    Chen, Xilin
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [29] Few-Shot Learning for Medical Image Classification
    Cai, Aihua
    Hu, Wenxin
    Zheng, Jun
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2020, PT I, 2020, 12396 : 441 - 452
  • [30] An Image Enhancement Method for Few-shot Classification
    Wu, Benze
    Wu, Yirui
    Wan, Shaohua
    [J]. 2021 IEEE 19TH INTERNATIONAL CONFERENCE ON EMBEDDED AND UBIQUITOUS COMPUTING (EUC 2021), 2021, : 159 - 165