y-Tuning: an efficient tuning paradigm for large-scale pre-trained models via label representation learning

被引:0
|
作者
Liu, Yitao [1 ]
An, Chenxin [1 ]
Qiu, Xipeng [1 ]
机构
[1] Fudan Univ, Sch Comp Sci, Shanghai 200433, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
pre-trained model; lightweight fine-tuning paradigms; label representation;
D O I
10.1007/s11704-023-3131-8
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With current success of large-scale pre-trained models (PTMs), how efficiently adapting PTMs to downstream tasks has attracted tremendous attention, especially for PTMs with billions of parameters. Previous work focuses on designing parameter-efficient tuning paradigms but needs to save and compute the gradient of the whole computational graph. In this paper, we propose y-Tuning, an efficient yet effective paradigm to adapt frozen large-scale PTMs to specific downstream tasks. y-Tuning learns dense representations for labels y defined in a given task and aligns them to fixed feature representation. Without computing the gradients of text encoder at training phrase, y-Tuning is not only parameter-efficient but also training-efficient. Experimental results show that for DeBERTa(XXL) with 1.6 billion parameters, y-Tuning achieves performance more than 96% of full fine-tuning on GLUE Benchmark with only 2% tunable parameters and much fewer training costs.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] Exploring the Application of Large-Scale Pre-Trained Models on Adverse Weather Removal
    Tan, Zhentao
    Wu, Yue
    Liu, Qiankun
    Chu, Qi
    Lu, Le
    Ye, Jieping
    Yu, Nenghai
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 1683 - 1698
  • [22] Large-scale Multi-modal Pre-trained Models: A Comprehensive Survey
    Wang, Xiao
    Chen, Guangyao
    Qian, Guangwu
    Gao, Pengcheng
    Wei, Xiao-Yong
    Wang, Yaowei
    Tian, Yonghong
    Gao, Wen
    MACHINE INTELLIGENCE RESEARCH, 2023, 20 (04) : 447 - 482
  • [23] Large-scale Multi-modal Pre-trained Models: A Comprehensive Survey
    Xiao Wang
    Guangyao Chen
    Guangwu Qian
    Pengcheng Gao
    Xiao-Yong Wei
    Yaowei Wang
    Yonghong Tian
    Wen Gao
    Machine Intelligence Research, 2023, 20 : 447 - 482
  • [24] FASTERMOE: Modeling and Optimizing Training of Large-Scale Dynamic Pre-Trained Models
    He, Jiaao
    Zhai, Jidong
    Antunes, Tiago
    Wang, Haojie
    Luo, Fuwen
    Shi, Shangfeng
    Li, Qin
    PPOPP'22: PROCEEDINGS OF THE 27TH ACM SIGPLAN SYMPOSIUM ON PRINCIPLES AND PRACTICE OF PARALLEL PROGRAMMING, 2022, : 120 - 134
  • [25] Alternating Recurrent Dialog Model with Large-scale Pre-trained Language Models
    Wu, Qingyang
    Zhang, Yichi
    Li, Yu
    Yu, Zhou
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 1292 - 1301
  • [26] PPT: Pre-trained Prompt Tuning for Few-shot Learning
    Gu, Yuxian
    Han, Xu
    Liu, Zhiyuan
    Huang, Minlie
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 8410 - 8423
  • [27] Large-Scale Relation Learning for Question Answering over Knowledge Bases with Pre-trained Language Models
    Yam, Yuanmeng
    Li, Rumei
    Wang, Sirui
    Zhang, Hongzhi
    Zan, Daoguang
    Zhang, Fuzheng
    Wu, Wei
    Xu, Weiran
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 3653 - 3660
  • [28] Hadamard Adapter: An Extreme Parameter-Efficient Adapter Tuning Method for Pre-trained Language Models
    Chen, Yuyan
    Fu, Qiang
    Fan, Ge
    Du, Lun
    Lou, Jian-Guang
    Han, Shi
    Zhang, Dongmei
    Li, Zhixu
    Xiao, Yanghua
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 276 - 285
  • [29] Personalised soft prompt tuning in pre-trained language models: Bridging multitask transfer learning and crowdsourcing learning
    Tian, Zeshu
    Zhang, Hongli
    Wang, Yan
    Knowledge-Based Systems, 2024, 305
  • [30] An Empirical Study of Parameter-Efficient Fine-Tuning Methods for Pre-trained Code Models
    Liu, Jiaxing
    Sha, Chaofeng
    Peng, Xin
    2023 38TH IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING, ASE, 2023, : 397 - 408