FD-Align: Feature Discrimination Alignment for Fine-tuning Pre-Trained Models in Few-Shot Learning

被引:0
|
作者
Song, Kun [1 ]
Ma, Huimin [1 ]
Zou, Bochao [1 ]
Zhang, Huishuai [3 ]
Huang, Weiran [2 ]
机构
[1] Univ Sci & Technol Beijing, SCCE, Beijing, Peoples R China
[2] Shanghai Jiao Tong Univ, SEIEE, Qing Yuan Res Inst, Shanghai, Peoples R China
[3] Microsoft Res Asia, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Due to the limited availability of data, existing few-shot learning methods trained from scratch fail to achieve satisfactory performance. In contrast, large-scale pre-trained models such as CLIP demonstrate remarkable few-shot and zero-shot capabilities. To enhance the performance of pre-trained models for downstream tasks, fine-tuning the model on downstream data is frequently necessary. However, fine-tuning the pre-trained model leads to a decrease in its generalizability in the presence of distribution shift, while the limited number of samples in few-shot learning makes the model highly susceptible to overfitting. Consequently, existing methods for fine-tuning few-shot learning primarily focus on fine-tuning the model's classification head or introducing additional structure. In this paper, we introduce a fine-tuning approach termed Feature Discrimination Alignment (FD-Align). Our method aims to bolster the model's generalizability by preserving the consistency of spurious features across the fine-tuning process. Extensive experimental results validate the efficacy of our approach for both ID and OOD tasks. Once fine-tuned, the model can seamlessly integrate with existing methods, leading to performance improvements. Our code could be found in https://github.com/skingorz/FD-Align.
引用
收藏
页数:14
相关论文
共 50 条
  • [31] Few-Shot Intent Detection via Contrastive Pre-Training and Fine-Tuning
    Zhang, Jian-Guo
    Bui, Trung
    Yoon, Seunghyun
    Chen, Xiang
    Liu, Zhiwei
    Xia, Congying
    Tran, Quan Hung
    Chang, Walter
    Yu, Philip
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 1906 - 1912
  • [32] EFTNet: an efficient fine-tuning method for few-shot segmentation
    Li, Jiaguang
    Wang, Yubo
    Gao, Zihan
    Wei, Ying
    APPLIED INTELLIGENCE, 2024, 54 (19) : 9488 - 9507
  • [33] Overcoming Catastrophic Forgetting for Fine-Tuning Pre-trained GANs
    Zhang, Zeren
    Li, Xingjian
    Hong, Tao
    Wang, Tianyang
    Ma, Jinwen
    Xiong, Haoyi
    Xu, Cheng-Zhong
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, ECML PKDD 2023, PT V, 2023, 14173 : 293 - 308
  • [34] Partial Is Better Than All: Revisiting Fine-tuning Strategy for Few-shot Learning
    Shen, Zhiqiang
    Liu, Zechun
    Qin, Jie
    Savvides, Marios
    Cheng, Kwang-Ting
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 9594 - 9602
  • [35] Waste Classification by Fine-Tuning Pre-trained CNN and GAN
    Alsabei, Amani
    Alsayed, Ashwaq
    Alzahrani, Manar
    Al-Shareef, Sarah
    INTERNATIONAL JOURNAL OF COMPUTER SCIENCE AND NETWORK SECURITY, 2021, 21 (08): : 65 - 70
  • [36] Pre-trained Vision and Language Transformers Are Few-Shot Incremental Learners
    Park, Keon-Hee
    Song, Kyungwoo
    Park, Gyeong-Moon
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 23881 - 23890
  • [37] COMPARING THE EFFICACY OF FINE-TUNING AND META-LEARNING FOR FEW-SHOT POLICY IMITATION
    Patacchiola, Massimiliano
    Sun, Mingfei
    Hofmann, Katja
    Turner, Richard E.
    CONFERENCE ON LIFELONG LEARNING AGENTS, VOL 232, 2023, 232 : 878 - 908
  • [38] Better Few-Shot Text Classification with Pre-trained Language Model
    Chen, Zheng
    Zhang, Yunchen
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT II, 2021, 12892 : 537 - 548
  • [39] Few-Sample Named Entity Recognition for Security Vulnerability Reports by Fine-Tuning Pre-trained Language Models
    Yang, Guanqun
    Dineen, Shay
    Lin, Zhipeng
    Liu, Xueqing
    DEPLOYABLE MACHINE LEARNING FOR SECURITY DEFENSE, MLHAT 2021, 2021, 1482 : 55 - 78
  • [40] Few-shot medical relation extraction via prompt tuning enhanced pre-trained language model
    He, Guoxiu
    Huang, Chen
    NEUROCOMPUTING, 2025, 633