The Power of Scale for Parameter-Efficient Prompt Tuning

被引:0
|
作者
Lester, Brian [1 ]
Al-Rfou, Rami [1 ]
Constant, Noah [1 ]
机构
[1] Google Res, Mountain View, CA 94043 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this work, we explore "prompt tuning," a simple yet effective mechanism for learning "soft prompts" to condition frozen language models to perform specific downstream tasks. Unlike the discrete text prompts used by GPT-3, soft prompts are learned through back-propagation and can be tuned to incorporate signals from any number of labeled examples. Our end-to-end learned approach outperforms GPT-3's few-shot learning by a large margin. More remarkably, through ablations on model size using T5, we show that prompt tuning becomes more competitive with scale: as models exceed billions of parameters, our method "closes the gap" and matches the strong performance of model tuning (where all model weights are tuned). This finding is especially relevant because large models are costly to share and serve and the ability to reuse one frozen model for multiple downstream tasks can ease this burden. Our method can be seen as a simplification of the recently proposed "prefix tuning" of Li and Liang (2021) and we provide a comparison to this and other similar approaches. Finally, we show that conditioning a frozen model with soft prompts confers benefits in robustness to domain transfer and enables efficient "prompt ensembling." We release code and model checkpoints to reproduce our experiments.(1)
引用
收藏
页码:3045 / 3059
页数:15
相关论文
共 50 条
  • [1] Prompt tuning for parameter-efficient medical image segmentation
    Fischer, Marc
    Bartler, Alexander
    Yang, Bin
    MEDICAL IMAGE ANALYSIS, 2024, 91
  • [2] Parameter-Efficient Low-Resource Dialogue State Tracking by Prompt Tuning
    Ma, Mingyu Derek
    Kao, Jiun-Yu
    Gao, Shuyang
    Gupta, Arpit
    Jin, Di
    Chung, Tagyoung
    Peng, Nanyun
    INTERSPEECH 2023, 2023, : 4653 - 4657
  • [3] Parameter-Efficient Prompt Tuning Makes Generalized and Calibrated Neural Text Retrievers
    Tam, Weng Lam
    Liu, Xiao
    Ji, Kaixuan
    Xue, Lilong
    Zhang, Xingjian
    Dong, Yuxiao
    Lin, Jiahua
    Hu, Maodi
    Tang, Jie
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 13117 - 13130
  • [4] Federated Learning of Large Language Models with Parameter-Efficient Prompt Tuning and Adaptive Optimization
    Che, Tianshi
    Liu, Ji
    Zhou, Yang
    Ren, Jiaxiang
    Zhou, Jiwen
    Sheng, Victor S.
    Dai, Huaiyu
    Dou, Dejing
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING, EMNLP 2023, 2023, : 7871 - 7888
  • [5] Parameter-efficient Tuning of Large-scale Multimodal Foundation Model
    Wang, Haixin
    Yang, Xinlong
    Chang, Jianlong
    Jin, Dian
    Sun, Jinan
    Zhang, Shikun
    Luo, Xiao
    Tian, Qi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [6] Attention Prompt Tuning: Parameter-efficient Adaptation of Pre-trained Models for Action Recognition
    Bandara, Wele Gedara Chaminda
    Patel, Vishal M.
    2024 IEEE 18TH INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION, FG 2024, 2024,
  • [7] Dynamic Adapter Meets Prompt Tuning: Parameter-Efficient Transfer Learning for Point Cloud Analysis
    Zhou, Xin
    Liang, Dingkang
    Xu, Wei
    Zhu, Xingkui
    Xu, Yihan
    Zou, Zhikang
    Bai, Xiang
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 14707 - 14717
  • [8] Parameter-Efficient Tuning with Special Token Adaptation
    Yang, Xiaocong
    Huang, James Y.
    Zhou, Wenxuan
    Chen, Muhao
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 865 - 872
  • [9] On the Effectiveness of Parameter-Efficient Fine-Tuning
    Fu, Zihao
    Yang, Haoran
    So, Anthony Man-Cho
    Lam, Wai
    Bing, Lidong
    Collier, Nigel
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11, 2023, : 12799 - 12807
  • [10] Stochastic Bridges as Effective Regularizers for Parameter-Efficient Tuning
    Chen, Weize
    Han, Xu
    Lin, Yankai
    Liu, Zhiyuan
    Sun, Maosong
    Zhou, Jie
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 10400 - 10420