Meta-Adapters: Parameter Efficient Few-shot Fine-tuning through Meta-Learning

被引:0
|
作者
Bansal, Trapit [1 ]
Alzubi, Salaheddin [1 ]
Wang, Tong [2 ]
Lee, Jay-Yoon [1 ]
McCallum, Andrew [1 ]
机构
[1] Univ Massachusetts, Amherst, MA 01003 USA
[2] Microsoft Res, Montreal, PQ, Canada
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Consistent improvements in the representational capacity of large pre-trained transformers has made it increasingly viable to serve these models as shared priors that can be fine-tuned on a large number of downstream tasks. However, fine-tuning the entire model for every task of interest makes a copy of all the model parameters, rendering such scenarios highly impractical. Recently introduced Adapter methods propose a promising alternative, one where only a small number of additional parameters are introduced per task specifically for fine-tuning. However, Adapters often require large amounts of task-specific data for good performance and don't work well in data-scarce few-shot scenarios. In this paper, we approach parameter-efficient fine-tuning in few-shot settings from a meta-learning perspective. We introduce Meta-Adapters, which are small blocks of meta-learned adapter layers inserted in a pre-trained model that re-purpose a frozen pre-trained model into a parameter-efficient few-shot learner. Meta-Adapters perform competitively with state-of-the-art few-shot learning methods that require full fine-tuning, while only fine-tuning 0.6% of the parameters. We evaluate Meta-Adapters along with multiple transfer learning baselines on an evaluation suite of 17 classification tasks and find that they improve few-shot accuracy by a large margin over competitive parameter-efficient methods, while requiring significantly lesser parameters for fine-tuning. Moreover, when comparing few-shot prompting of GPT-3 against few-shot fine-tuning with Meta-Adapters, we find that Meta-Adapters perform competitively while working with pre-trained transformers that are many orders of magnitude (1590x) smaller in size than GPT-3.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] COMPARING THE EFFICACY OF FINE-TUNING AND META-LEARNING FOR FEW-SHOT POLICY IMITATION
    Patacchiola, Massimiliano
    Sun, Mingfei
    Hofmann, Katja
    Turner, Richard E.
    CONFERENCE ON LIFELONG LEARNING AGENTS, VOL 232, 2023, 232 : 878 - 908
  • [2] Strong Baselines for Parameter-Efficient Few-Shot Fine-Tuning
    Basu, Samyadeep
    Hu, Shell
    Massiceti, Daniela
    Feizi, Soheil
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 10, 2024, : 11024 - 11031
  • [3] Unsupervised meta-learning for few-shot learning
    Xu, Hui
    Wang, Jiaxing
    Li, Hao
    Ouyang, Deqiang
    Shao, Jie
    PATTERN RECOGNITION, 2021, 116
  • [4] Few-Shot Text Classification with an Efficient Prompt Tuning Method in Meta-Learning Framework
    Lv, Xiaobao
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2024, 38 (03)
  • [5] Meta-Learning for Few-Shot NMT Adaptation
    Sharaf, Amr
    Hassan, Hany
    Daume, Hal, III
    NEURAL GENERATION AND TRANSLATION, 2020, : 43 - 53
  • [6] Fair Meta-Learning For Few-Shot Classification
    Zhao, Chen
    Li, Changbin
    Li, Jincheng
    Chen, Feng
    11TH IEEE INTERNATIONAL CONFERENCE ON KNOWLEDGE GRAPH (ICKG 2020), 2020, : 275 - 282
  • [7] Adaptive fine-tuning strategy for few-shot learning
    Zhuang, Xinkai
    Shao, Mingwen
    Gao, Wei
    Yang, Jianxin
    JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (06)
  • [8] Task Agnostic Meta-Learning for Few-Shot Learning
    Jamal, Muhammad Abdullah
    Qi, Guo-Jun
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 11711 - 11719
  • [9] META-LEARNING WITH ATTENTION FOR IMPROVED FEW-SHOT LEARNING
    Hou, Zejiang
    Walid, Anwar
    Kung, Sun-Yuan
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 2725 - 2729
  • [10] Few-shot classification via efficient meta-learning with hybrid optimization
    Jia, Jinfang
    Feng, Xiang
    Yu, Huiqun
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 127