Integrating Task Specific Information into Pretrained Language Models for Low Resource Fine Tuning

被引:0
|
作者
Wang, Rui [1 ]
Si, Shijing [1 ]
Wang, Guoyin [1 ,2 ]
Zhang, Lei [3 ]
Carin, Lawrence [1 ]
Henao, Ricardo [1 ]
机构
[1] Duke Univ, Durham, NC 27706 USA
[2] Amazon Alexa AI, Cambridge, MA USA
[3] Fidel Investments, Raleigh, NC USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Pretrained Language Models (PLMs) have improved the performance of natural language understanding in recent years. Such models are pretrained on large corpora, which encode the general prior knowledge of natural languages but are agnostic to information characteristic of downstream tasks. This often results in overfitting when fine-tuned with low resource datasets where task-specific information is limited. In this paper, we integrate label information as a task-specific prior into the self-attention component of pretrained BERT models. Experiments on several benchmarks and real-word datasets suggest that the proposed approach can largely improve the performance of pretrained models when finetuning with small datasets. The code repository is released in https://github.com/RayWangWR/BERT_label_embedding.
引用
收藏
页数:6
相关论文
共 50 条
  • [1] Task-guided Disentangled Tuning for Pretrained Language Models
    Zeng, Jiali
    Jiang, Yufan
    Wu, Shuangzhi
    Yin, Yongjing
    Li, Mu
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 3126 - 3137
  • [2] CONVFIT: Conversational Fine-Tuning of Pretrained Language Models
    Vulic, Ivan
    Su, Pei-Hao
    Coope, Sam
    Gerz, Daniela
    Budzianowski, Pawel
    Casanueva, Inigo
    Mrksic, Nikola
    Wen, Tsung-Hsien
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 1151 - 1168
  • [3] DN at SemEval-2023 Task 12: Low-Resource Language Text Classification via Multilingual Pretrained Language Model Fine-tuning
    Daniil, Homskiy
    Narek, Maloyan
    17TH INTERNATIONAL WORKSHOP ON SEMANTIC EVALUATION, SEMEVAL-2023, 2023, : 1537 - 1541
  • [4] Fine Tuning Language Models: A Tale of Two Low-Resource Languages
    Rosel OidaOnesa
    Melvin ABallera
    Data Intelligence, 2024, 6 (04) : 946 - 967
  • [5] Low-resource Taxonomy Enrichment with Pretrained Language Models
    Takeoka, Kunihiro
    Akimoto, Kosuke
    Oyamada, Masafumi
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 2747 - 2758
  • [6] Recall and Learn: Fine-tuning Deep Pretrained Language Models with Less Forgetting
    Chen, Sanyuan
    Hou, Yutai
    Cui, Yiming
    Che, Wanxiang
    Liu, Ting
    Yu, Xiangzhan
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 7870 - 7881
  • [7] Toward Low-Resource Languages Machine Translation: A Language-Specific Fine-Tuning With LoRA for Specialized Large Language Models
    Liang, Xiao
    Khaw, Yen-Min Jasmina
    Liew, Soung-Yue
    Tan, Tien-Ping
    Qin, Donghong
    IEEE ACCESS, 2025, 13 : 46616 - 46626
  • [8] Fine-Tuning Pretrained Language Models to Enhance Dialogue Summarization in Customer Service Centers
    Yun, Jiseon
    Sohn, Jae Eui
    Kyeong, Sunghyon
    PROCEEDINGS OF THE 4TH ACM INTERNATIONAL CONFERENCE ON AI IN FINANCE, ICAIF 2023, 2023, : 365 - 373
  • [9] Noise-Robust Fine-Tuning of Pretrained Language Models via External Guidance
    Wang, Song
    Tan, Zhen
    Guo, Ruocheng
    Li, Jundong
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 12528 - 12540
  • [10] Vietnamese Sentiment Analysis: An Overview and Comparative Study of Fine-tuning Pretrained Language Models
    Dang Van Thin
    Duong Ngoc Hao
    Ngan Luu-Thuy Nguyen
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (06)