Task-guided Disentangled Tuning for Pretrained Language Models

被引:0
|
作者
Zeng, Jiali [1 ]
Jiang, Yufan [1 ]
Wu, Shuangzhi [1 ]
Yin, Yongjing [2 ]
Li, Mu [1 ]
机构
[1] Tencent Cloud Xiaowei, Beijing, Peoples R China
[2] Zhejiang Univ, Westlake Univ, Hangzhou, Zhejiang, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pretrained language models (PLMs) trained on large-scale unlabeled corpus are typically finetuned on task-specific downstream datasets, which have produced state-of-the-art results on various NLP tasks. However, the data discrepancy issue in domain and scale makes finetuning fail to efficiently capture task-specific patterns, especially in the low data regime. To address this issue, we propose Task-guided Disentangled Tuning (TDT) for PLMs, which enhances the generalization of representations by disentangling task-relevant signals from the entangled representations. For a given task, we introduce a learnable confidence model to detect indicative guidance from context, and further propose a disentangled regularization to mitigate the over-reliance problem. Experimental results on GLUE and CLUE benchmarks show that TDT gives consistently better results than fine-tuning with different PLMs, and extensive analysis demonstrates the effectiveness and robustness of our method. Code is available at https://github.com/lemon0830/TDT.
引用
收藏
页码:3126 / 3137
页数:12
相关论文
共 50 条
  • [1] Integrating Task Specific Information into Pretrained Language Models for Low Resource Fine Tuning
    Wang, Rui
    Si, Shijing
    Wang, Guoyin
    Zhang, Lei
    Carin, Lawrence
    Henao, Ricardo
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020,
  • [2] Task-guided IRL in POMDPs that scales
    Djeumou, Franck
    Ellis, Christian
    Cubuktepe, Murat
    Lennon, Craig
    Topcu, Ufuk
    ARTIFICIAL INTELLIGENCE, 2023, 317
  • [3] CONVFIT: Conversational Fine-Tuning of Pretrained Language Models
    Vulic, Ivan
    Su, Pei-Hao
    Coope, Sam
    Gerz, Daniela
    Budzianowski, Pawel
    Casanueva, Inigo
    Mrksic, Nikola
    Wen, Tsung-Hsien
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 1151 - 1168
  • [4] Task-Guided Pair Embedding in Heterogeneous Network
    Park, Chanyoung
    Kim, Donghyun
    Zhu, Qi
    Han, Jiawei
    Yu, Hwanjo
    PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT (CIKM '19), 2019, : 489 - 498
  • [5] A Framework for Task-Guided Virtual Machine Live Migration
    Lin, Cho-Chin
    Kuo, Yuan-Han
    INTERNET AND DISTRIBUTED COMPUTING SYSTEMS, 2018, 11226 : 40 - 51
  • [6] Task-guided selection of the dual neural pathways for reading
    Nakamura, Kimihiro
    Hara, Nobuko
    Kouider, Sid
    Takayama, Yoshihiro
    Hanajima, Ritsuko
    Sakai, Katsuyuki
    Ugawa, Yoshikazu
    NEURON, 2006, 52 (03) : 557 - 564
  • [7] A Survey of Pretrained Language Models
    Sun, Kaili
    Luo, Xudong
    Luo, Michael Y.
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT II, 2022, 13369 : 442 - 456
  • [8] ProtoDiff: Learning to Learn Prototypical Networks by Task-Guided Diffusion
    Du, Yingjun
    Xiao, Zehao
    Liao, Shencai
    Snoek, Cees G. M.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [9] Task-Guided Context-Path Embedding in Temporal Heterogeneous Networks
    Hu, Qian
    Lin, Fan
    Wang, Beizhan
    Li, Chunyan
    IEEE ACCESS, 2020, 8 : 205170 - 205180
  • [10] Task Residual for Tuning Vision-Language Models
    Yu, Tao
    Lu, Zhihe
    Jin, Xin
    Chen, Zhibo
    Wang, Xinchao
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 10899 - 10909