Explore pretraining for few-shot learning

被引:0
|
作者
Li, Yan [1 ]
Huang, Jinjie [1 ]
机构
[1] Harbin Univ Sci & Technol, Inst Automat, Harbin 150006, Heilongjiang, Peoples R China
关键词
Computer vision; Deep learning; Image classification; Few-shot learning;
D O I
10.1007/s11042-023-15223-1
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Few-shot learning aims to learn to classify new categories with a few samples. Pretraining the model on the base class can improve the performance of the model on the new category. To further improve the pretraining performance of the model on the base class, we propose a two-stage model pretraining method. In the first stage, we conduct Simsiam contrastive learning pretraining, which can help the model learn invariant knowledge. In the second stage, we conduct multi-task pretraining for general classification tasks and rotation prediction tasks, which can help the model learn the equivalent knowledge. Two pretraining stages can significantly enhance the model's capacity to learn new categories and enhance the effectiveness of few-shot categorization. Experiments show that our method achieves State-of-the-art few-shot classification performance on the mini-ImageNet and FC100 datasets for 1-shot and 5-shot tasks.
引用
收藏
页码:4691 / 4702
页数:12
相关论文
共 50 条
  • [1] Explore pretraining for few-shot learning
    Yan Li
    Jinjie Huang
    [J]. Multimedia Tools and Applications, 2024, 83 : 4691 - 4702
  • [2] ExPT: Synthetic Pretraining for Few-Shot Experimental Design
    Nguyen, Tung
    Agrawal, Sudhanshu
    Grover, Aditya
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [3] Few-Shot Question Answering by Pretraining Span Selection
    Ram, Ori
    Kirstain, Yuval
    Berant, Jonathan
    Globerson, Amir
    Levy, Omer
    [J]. 59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 3066 - 3079
  • [4] Few-Shot Few-Shot Learning and the role of Spatial Attention
    Lifchitz, Yann
    Avrithis, Yannis
    Picard, Sylvaine
    [J]. 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 2693 - 2700
  • [5] Explore Across-Dimensional Feature Correlations for Few-Shot Learning
    Li, Tianyuan
    Cao, Wenming
    [J]. PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT III, 2024, 14427 : 151 - 163
  • [6] Few-shot Relational Reasoning via Connection Subgraph Pretraining
    Huang, Qian
    Ren, Hongyu
    Leskovec, Jure
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [7] DFCP: Few-Shot DeepFake Detection via Contrastive Pretraining
    Zou, Bo
    Yang, Chao
    Guan, Jiazhi
    Quan, Chengbin
    Zhao, Youjian
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 2303 - 2308
  • [8] Variational Few-Shot Learning
    Zhang, Jian
    Zhao, Chenglong
    Ni, Bingbing
    Xu, Minghao
    Yang, Xiaokang
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 1685 - 1694
  • [9] Survey on Few-shot Learning
    Zhao, Kai-Lin
    Jin, Xiao-Long
    Wang, Yuan-Zhuo
    [J]. Ruan Jian Xue Bao/Journal of Software, 2021, 32 (02): : 349 - 369
  • [10] Federated Few-shot Learning
    Wang, Song
    Fu, Xingbo
    Ding, Kaize
    Chen, Chen
    Chen, Huiyuan
    Li, Jundong
    [J]. PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 2374 - 2385