Meta-Adapters: Parameter Efficient Few-shot Fine-tuning through Meta-Learning

被引:0
|
作者
Bansal, Trapit [1 ]
Alzubi, Salaheddin [1 ]
Wang, Tong [2 ]
Lee, Jay-Yoon [1 ]
McCallum, Andrew [1 ]
机构
[1] Univ Massachusetts, Amherst, MA 01003 USA
[2] Microsoft Res, Montreal, PQ, Canada
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Consistent improvements in the representational capacity of large pre-trained transformers has made it increasingly viable to serve these models as shared priors that can be fine-tuned on a large number of downstream tasks. However, fine-tuning the entire model for every task of interest makes a copy of all the model parameters, rendering such scenarios highly impractical. Recently introduced Adapter methods propose a promising alternative, one where only a small number of additional parameters are introduced per task specifically for fine-tuning. However, Adapters often require large amounts of task-specific data for good performance and don't work well in data-scarce few-shot scenarios. In this paper, we approach parameter-efficient fine-tuning in few-shot settings from a meta-learning perspective. We introduce Meta-Adapters, which are small blocks of meta-learned adapter layers inserted in a pre-trained model that re-purpose a frozen pre-trained model into a parameter-efficient few-shot learner. Meta-Adapters perform competitively with state-of-the-art few-shot learning methods that require full fine-tuning, while only fine-tuning 0.6% of the parameters. We evaluate Meta-Adapters along with multiple transfer learning baselines on an evaluation suite of 17 classification tasks and find that they improve few-shot accuracy by a large margin over competitive parameter-efficient methods, while requiring significantly lesser parameters for fine-tuning. Moreover, when comparing few-shot prompting of GPT-3 against few-shot fine-tuning with Meta-Adapters, we find that Meta-Adapters perform competitively while working with pre-trained transformers that are many orders of magnitude (1590x) smaller in size than GPT-3.
引用
收藏
页数:18
相关论文
共 50 条
  • [41] A concise review of recent few-shot meta-learning methods
    Li, Xiaoxu
    Sun, Zhuo
    Xue, Jing-Hao
    Ma, Zhanyu
    NEUROCOMPUTING, 2021, 456 : 463 - 468
  • [42] Few-Shot Classification Based on Sparse Dictionary Meta-Learning
    Jiang, Zuo
    Wang, Yuan
    Tang, Yi
    MATHEMATICS, 2024, 12 (19)
  • [43] Prototype Bayesian Meta-Learning for Few-Shot Image Classification
    Fu, Meijun
    Wang, Xiaomin
    Wang, Jun
    Yi, Zhang
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 15
  • [44] MetaDelta: A Meta-Learning System for Few-shot Image Classification
    Chen, Yudong
    Guan, Chaoyu
    Wei, Zhikun
    Wang, Xin
    Zhu, Wenwu
    AAAI WORKSHOP ON META-LEARNING AND METADL CHALLENGE, VOL 140, 2021, 140 : 17 - 28
  • [45] Few-shot and meta-learning methods for image understanding: a survey
    He, Kai
    Pu, Nan
    Lao, Mingrui
    Lew, Michael S. S.
    INTERNATIONAL JOURNAL OF MULTIMEDIA INFORMATION RETRIEVAL, 2023, 12 (02)
  • [46] Weakly Supervised Few-Shot Segmentation via Meta-Learning
    Gama, Pedro H. T.
    Oliveira, Hugo
    Marcato Jr, Jose
    dos Santos, Jefersson A.
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 1784 - 1797
  • [47] Few-shot time series forecasting in a meta-learning framework
    Ma P.
    Ni Z.
    Ma, Ping (1533321767@qq.com), 1600, IOS Press BV (46): : 8903 - 8916
  • [48] Meta-Learning for Multi-Label Few-Shot Classification
    Simon, Christian
    Koniusz, Piotr
    Harandi, Mehrtash
    2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), 2022, : 346 - 355
  • [49] Differentiable Meta-Learning Model for Few-Shot Semantic Segmentation
    Tian, Pinzhuo
    Wu, Zhangkai
    Qi, Lei
    Wang, Lei
    Shi, Yinghuan
    Gao, Yang
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 12087 - 12094
  • [50] Few-shot and meta-learning methods for image understanding: a survey
    Kai He
    Nan Pu
    Mingrui Lao
    Michael S. Lew
    International Journal of Multimedia Information Retrieval, 2023, 12