On the Effectiveness of Adapter-based Tuning for Pretrained Language Model Adaptation

被引:0
|
作者
He, Ruidan [1 ]
Liu, Linlin [1 ,2 ]
Ye, Hai [3 ]
Tan, Qingyu [1 ,3 ]
Ding, Bosheng [1 ,2 ]
Cheng, Liying [1 ,4 ]
Low, Jia-Wei [1 ,2 ]
Bing, Lidong [1 ]
Si, Luo [1 ]
机构
[1] Alibaba Grp, DAMO Acad, Hangzhou, Peoples R China
[2] Nanyang Technol Univ, Singapore, Singapore
[3] Natl Univ Singapore, Singapore, Singapore
[4] Singapore Univ Technol & Design, Singapore, Singapore
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adapter-based tuning has recently arisen as an alternative to fine-tuning. It works by adding light-weight adapter modules to a pretrained language model (PrLM) and only updating the parameters of adapter modules when learning on a downstream task. As such, it adds only a few trainable parameters per new task, allowing a high degree of parameter sharing. Prior studies have shown that adapter-based tuning often achieves comparable results to finetuning. However, existing work only focuses on the parameter-efficient aspect of adapterbased tuning while lacking further investigation on its effectiveness. In this paper, we study the latter. We first show that adapterbased tuning better mitigates forgetting issues than fine-tuning since it yields representations with less deviation from those generated by the initial PrLM. We then empirically compare the two tuning methods on several downstream NLP tasks and settings. We demonstrate that 1) adapter-based tuning outperforms fine-tuning on low-resource and cross-lingual tasks; 2) it is more robust to overfitting and less sensitive to changes in learning rates.
引用
收藏
页码:2208 / 2222
页数:15
相关论文
共 50 条
  • [41] DagoBERT: Generating Derivational Morphology with a Pretrained Language Model
    Hofmannt, Valentin
    Pierrehumbertt, Janet B.
    Schiitzet, Hinrich
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 3848 - 3861
  • [42] INTEGRATING PRETRAINED LANGUAGE MODEL FOR DIALOGUE POLICY EVALUATION
    Wang, Hongru
    Wang, Huimin
    Wang, Zezhong
    Wong, Kam-Fai
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 6692 - 6696
  • [43] BatteryBERT: A Pretrained Language Model for Battery Database Enhancement
    Huang, Shu
    Cole, Jacqueline M.
    JOURNAL OF CHEMICAL INFORMATION AND MODELING, 2022, 62 (24) : 6365 - 6377
  • [44] Search and Classification Based Language Model Adaptation
    Shi, Qin
    Chu, Stephen M.
    Liu, Wen
    Kuo, Hong-Kwang
    Liu, Yi
    Qin, Yong
    INTERSPEECH 2008: 9TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2008, VOLS 1-5, 2008, : 1578 - 1581
  • [45] MAD-X: An Adapter-Based Framework for Multi-Task Cross-Lingual Transfer
    Pfeiffer, Jonas
    Vulic, Ivan
    Gurevych, Iryna
    Ruder, Sebastian
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 7654 - 7673
  • [46] VisualGPT: Data-efficient Adaptation of Pretrained Language Models for Image Captioning
    Chen, Jun
    Guo, Han
    Yi, Kai
    Li, Boyang
    Elhoseiny, Mohamed
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 18009 - 18019
  • [47] A Survey on Model Compression and Acceleration for Pretrained Language Models
    Xu, Canwen
    McAuley, Julian
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 9, 2023, : 10566 - 10575
  • [48] KRongBERT: Enhanced factorization-based morphological approach for the Korean pretrained language model
    Yu, Hyunwook
    Cho, Yejin
    Park, Geunchul
    Kim, Mucheol
    INFORMATION PROCESSING & MANAGEMENT, 2025, 62 (03)
  • [49] Comprehensive prediction and analysis of human protein essentiality based on a pretrained large language model
    Kang, Boming
    Fan, Rui
    Cui, Chunmei
    Cui, Qinghua
    NATURE COMPUTATIONAL SCIENCE, 2024, : 196 - 206
  • [50] Integrating Task Specific Information into Pretrained Language Models for Low Resource Fine Tuning
    Wang, Rui
    Si, Shijing
    Wang, Guoyin
    Zhang, Lei
    Carin, Lawrence
    Henao, Ricardo
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020,