On the Effectiveness of Adapter-based Tuning for Pretrained Language Model Adaptation

被引:0
|
作者
He, Ruidan [1 ]
Liu, Linlin [1 ,2 ]
Ye, Hai [3 ]
Tan, Qingyu [1 ,3 ]
Ding, Bosheng [1 ,2 ]
Cheng, Liying [1 ,4 ]
Low, Jia-Wei [1 ,2 ]
Bing, Lidong [1 ]
Si, Luo [1 ]
机构
[1] Alibaba Grp, DAMO Acad, Hangzhou, Peoples R China
[2] Nanyang Technol Univ, Singapore, Singapore
[3] Natl Univ Singapore, Singapore, Singapore
[4] Singapore Univ Technol & Design, Singapore, Singapore
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Adapter-based tuning has recently arisen as an alternative to fine-tuning. It works by adding light-weight adapter modules to a pretrained language model (PrLM) and only updating the parameters of adapter modules when learning on a downstream task. As such, it adds only a few trainable parameters per new task, allowing a high degree of parameter sharing. Prior studies have shown that adapter-based tuning often achieves comparable results to finetuning. However, existing work only focuses on the parameter-efficient aspect of adapterbased tuning while lacking further investigation on its effectiveness. In this paper, we study the latter. We first show that adapterbased tuning better mitigates forgetting issues than fine-tuning since it yields representations with less deviation from those generated by the initial PrLM. We then empirically compare the two tuning methods on several downstream NLP tasks and settings. We demonstrate that 1) adapter-based tuning outperforms fine-tuning on low-resource and cross-lingual tasks; 2) it is more robust to overfitting and less sensitive to changes in learning rates.
引用
收藏
页码:2208 / 2222
页数:15
相关论文
共 50 条
  • [31] Vison transformer adapter-based hyperbolic embeddings for multi-lesion segmentation in diabetic retinopathy
    Wang, Zijian
    Lu, Haimei
    Yan, Haixin
    Kan, Hongxing
    Jin, Li
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [32] Automatic Taxonomy Classification by Pretrained Language Model
    Kuwana, Ayato
    Oba, Atsushi
    Sawai, Ranto
    Paik, Incheon
    ELECTRONICS, 2021, 10 (21)
  • [33] Span-Based Nested Named Entity Recognition with Pretrained Language Model
    Liu, Chenxu
    Fan, Hongjie
    Liu, Junfei
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS (DASFAA 2021), PT II, 2021, 12682 : 620 - 628
  • [34] Adapter-based Safety Injection System for Prevention of Wrong Route and Wrong Patient Medication Errors
    Cho, Yong Chan
    Lee, Seung Ho
    Cho, Yang Hyun
    Choy, Young Bin
    JOURNAL OF KOREAN MEDICAL SCIENCE, 2017, 32 (12) : 1938 - +
  • [35] Vison transformer adapter-based hyperbolic embeddings for multi-lesion segmentation in diabetic retinopathy
    Zijian Wang
    Haimei Lu
    Haixin Yan
    Hongxing Kan
    Li Jin
    Scientific Reports, 13
  • [36] Language Recognition Based on Unsupervised Pretrained Models
    Yu, Haibin
    Zhao, Jing
    Yang, Song
    Wu, Zhongqin
    Nie, Yuting
    Zhang, Wei-Qiang
    INTERSPEECH 2021, 2021, : 3271 - 3275
  • [37] ATTEN-ADAPTER: A UNIFIED ATTENTION-BASED ADAPTER FOR EFFICIENT TUNING
    Li, Kaiwen
    Gu, Wenzhe
    Xue, Maixuan
    Xiao, Jiahua
    Shi, Dahu
    Wei, Xing
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 1265 - 1269
  • [38] Graph Adversarial Network with Bottleneck Adapter Tuning for Sign Language Production
    Yu, Chunpeng
    Liang, Jiajia
    Liao, Yihui
    Xie, Zhifeng
    Sheng, Bin
    ADVANCES IN COMPUTER GRAPHICS, CGI 2022, 2022, 13443 : 92 - 103
  • [39] An Empirical Study on Adaptive Inference for Pretrained Language Model
    Liu, Weijie
    Zhao, Xin
    Zhao, Zhe
    Ju, Qi
    Yang, Xuefeng
    Lu, Wei
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (08) : 4321 - 4331
  • [40] Hadamard Adapter: An Extreme Parameter-Efficient Adapter Tuning Method for Pre-trained Language Models
    Chen, Yuyan
    Fu, Qiang
    Fan, Ge
    Du, Lun
    Lou, Jian-Guang
    Han, Shi
    Zhang, Dongmei
    Li, Zhixu
    Xiao, Yanghua
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 276 - 285