MiDTD: A Simple and Effective Distillation Framework for Distantly Supervised Relation Extraction

被引:6
|
作者
Li, Rui [1 ]
Yang, Cheng [1 ]
Li, Tingwei [1 ]
Su, Sen [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Sch Comp Sci, Natl Pilot Software Engn Sch, 10 Xi Tucheng Rd, Beijing 100876, Peoples R China
基金
中国国家自然科学基金;
关键词
Natural language processing; NLP; knowledge distillation; distant supervision; neural network; multi-instance learning; label softening;
D O I
10.1145/3503917
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Relation extraction (RE), an important information extraction task, faced the great challenge brought by limited annotation data. To this end, distant supervision was proposed to automatically label RE data, and thus largely increased the number of annotated instances. Unfortunately, lots of noise relation annotations brought by automatic labeling become a new obstacle. Some recent studies have shown that the teacher-student framework of knowledge distillation can alleviate the interference of noise relation annotations via label softening. Nevertheless, we find that they still suffer from two problems: propagation of inaccurate dark knowledge and constraint of a unified distillation temperature. In this article, we propose a simple and effective Multi-instance Dynamic Temperature Distillation (MiDTD) framework, which is model-agnostic and mainly involves two modules: multi-instance target fusion (MiTF) and dynamic temperature regulation (DTR). MiTF combines the teacher's predictions for multiple sentences with the same entity pair to amend the inaccurate dark knowledge in each student's target. DTR allocates alterable distillation temperatures to different training instances to enable the softness of most student's targets to be regulated to a moderate range. In experiments, we construct three concrete MiDTD instantiations with BERT, PCNN, and BiLSTM-based RE models, and the distilled students significantly outperform their teachers and the state-of-the-art (SOTA) methods.
引用
收藏
页数:32
相关论文
共 50 条
  • [1] A Multi-teacher Knowledge Distillation Framework for Distantly Supervised Relation Extraction with Flexible Temperature
    Fei, Hongxiao
    Tan, Yangying
    Huang, Wenti
    Long, Jun
    Huang, Jincai
    Yang, Liu
    WEB AND BIG DATA, PT II, APWEB-WAIM 2023, 2024, 14332 : 103 - 116
  • [2] Noise Reduction Framework for Distantly Supervised Relation Extraction with Human in the Loop
    Zhang, Xinyuan
    Liu, Hongzhi
    Wu, Zhonghai
    PROCEEDINGS OF 2020 IEEE 10TH INTERNATIONAL CONFERENCE ON ELECTRONICS INFORMATION AND EMERGENCY COMMUNICATION (ICEIEC 2020), 2020, : 1 - 4
  • [3] HiCLRE: A Hierarchical Contrastive Learning Framework for Distantly Supervised Relation Extraction
    Li, Dongyang
    Zhang, Taolin
    Hu, Nan
    Wang, Chengyu
    He, Xiaofeng
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 2567 - 2578
  • [4] Distantly supervised relation extraction with a Meta-Relation enhanced Contrastive learning framework
    Chen, Chuanshu
    Hao, Shuang
    Liu, Jian
    NEUROCOMPUTING, 2025, 617
  • [5] PARE: A Simple and Strong Baseline for Monolingual and Multilingual Distantly Supervised Relation Extraction
    Rathore, Vipul
    Badola, Kartikeya
    Singla, Parag
    Mausam
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022): (SHORT PAPERS), VOL 2, 2022, : 340 - 354
  • [6] Distantly Supervised Relation Extraction in Federated Settings
    Sui, Dianbo
    Chen, Yubo
    Liu, Kang
    Zhao, Jun
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 569 - 583
  • [7] Survey on Distantly-Supervised Relation Extraction
    Yang S.-Z.
    Liu Y.-X.
    Zhang K.-W.
    Hong Y.
    Huang H.
    Jisuanji Xuebao/Chinese Journal of Computers, 2021, 44 (08): : 1636 - 1660
  • [8] Bias Modeling for Distantly Supervised Relation Extraction
    Xiang, Yang
    Zhang, Yaoyun
    Wang, Xiaolong
    Qin, Yang
    Han, Wenying
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2015, 2015
  • [9] Reinforcement Learning for Distantly Supervised Relation Extraction
    Sun, Tingting
    Zhang, Chunhong
    Ji, Yang
    Hu, Zheng
    IEEE ACCESS, 2019, 7 : 98023 - 98033
  • [10] MKDAT: Multi-Level Knowledge Distillation with Adaptive Temperature for Distantly Supervised Relation Extraction
    Long, Jun
    Yin, Zhuoying
    Han, Yan
    Huang, Wenti
    INFORMATION, 2024, 15 (07)