Task-guided Disentangled Tuning for Pretrained Language Models

被引:0
|
作者
Zeng, Jiali [1 ]
Jiang, Yufan [1 ]
Wu, Shuangzhi [1 ]
Yin, Yongjing [2 ]
Li, Mu [1 ]
机构
[1] Tencent Cloud Xiaowei, Beijing, Peoples R China
[2] Zhejiang Univ, Westlake Univ, Hangzhou, Zhejiang, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pretrained language models (PLMs) trained on large-scale unlabeled corpus are typically finetuned on task-specific downstream datasets, which have produced state-of-the-art results on various NLP tasks. However, the data discrepancy issue in domain and scale makes finetuning fail to efficiently capture task-specific patterns, especially in the low data regime. To address this issue, we propose Task-guided Disentangled Tuning (TDT) for PLMs, which enhances the generalization of representations by disentangling task-relevant signals from the entangled representations. For a given task, we introduce a learnable confidence model to detect indicative guidance from context, and further propose a disentangled regularization to mitigate the over-reliance problem. Experimental results on GLUE and CLUE benchmarks show that TDT gives consistently better results than fine-tuning with different PLMs, and extensive analysis demonstrates the effectiveness and robustness of our method. Code is available at https://github.com/lemon0830/TDT.
引用
收藏
页码:3126 / 3137
页数:12
相关论文
共 50 条
  • [41] Pretrained Models and Evaluation Data for the Khmer Language
    Shengyi Jiang
    Sihui Fu
    Nankai Lin
    Yingwen Fu
    Tsinghua Science and Technology, 2022, 27 (04) : 709 - 718
  • [42] Pretrained Language Models for Sequential Sentence Classification
    Cohan, Arman
    Beltagy, Iz
    King, Daniel
    Dalvi, Bhavana
    Weld, Daniel S.
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 3693 - 3699
  • [43] Pretrained Language Models for Text Generation: A Survey
    Li, Junyi
    Tang, Tianyi
    Zhao, Wayne Xin
    Wen, Ji-Rong
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 4492 - 4499
  • [44] Probing Pretrained Language Models with Hierarchy Properties
    Lovon-Melgarejo, Jesus
    Moreno, Jose G.
    Besancon, Romaric
    Ferret, Olivier
    Tamine, Lynda
    ADVANCES IN INFORMATION RETRIEVAL, ECIR 2024, PT II, 2024, 14609 : 126 - 142
  • [45] Probing Pretrained Language Models for Lexical Semantics
    Vulie, Ivan
    Ponti, Edoardo M.
    Litschko, Robert
    Glava, Goran
    Korhonen, Anna
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 7222 - 7240
  • [46] Auxiliary Task-Guided CycleGAN for Black-Box Model Domain Adaptation
    Essich, Michael
    Rehmann, Markus
    Curio, Cristobal
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 541 - 550
  • [47] Pretrained models and evaluation data for the Khmer language
    Jiang, Shengyi
    Fu, Sihui
    Lin, Nankai
    Fu, Yingwen
    TSINGHUA SCIENCE AND TECHNOLOGY, 2022, 27 (04) : 709 - 718
  • [48] Equi-Tuning: Group Equivariant Fine-Tuning of Pretrained Models
    Basu, Sourya
    Sattigeri, Prasanna
    Ramamurthy, Karthikeyan Natesan
    Chenthamarakshan, Vijil
    Varshney, Kush R.
    Varshney, Lav R.
    Das, Payel
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 6, 2023, : 6788 - 6796
  • [49] YNU-HPCC at SemEval-2022 Task 4: Finetuning Pretrained Language Models for Patronizing and Condescending Language Detection
    Bai, Wenqiang
    Wang, Jin
    Zhang, Xuejie
    PROCEEDINGS OF THE 16TH INTERNATIONAL WORKSHOP ON SEMANTIC EVALUATION, SEMEVAL-2022, 2022, : 454 - 458
  • [50] On the Effectiveness of Adapter-based Tuning for Pretrained Language Model Adaptation
    He, Ruidan
    Liu, Linlin
    Ye, Hai
    Tan, Qingyu
    Ding, Bosheng
    Cheng, Liying
    Low, Jia-Wei
    Bing, Lidong
    Si, Luo
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 2208 - 2222