MiDTD: A Simple and Effective Distillation Framework for Distantly Supervised Relation Extraction

被引:6
|
作者
Li, Rui [1 ]
Yang, Cheng [1 ]
Li, Tingwei [1 ]
Su, Sen [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Sch Comp Sci, Natl Pilot Software Engn Sch, 10 Xi Tucheng Rd, Beijing 100876, Peoples R China
基金
中国国家自然科学基金;
关键词
Natural language processing; NLP; knowledge distillation; distant supervision; neural network; multi-instance learning; label softening;
D O I
10.1145/3503917
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Relation extraction (RE), an important information extraction task, faced the great challenge brought by limited annotation data. To this end, distant supervision was proposed to automatically label RE data, and thus largely increased the number of annotated instances. Unfortunately, lots of noise relation annotations brought by automatic labeling become a new obstacle. Some recent studies have shown that the teacher-student framework of knowledge distillation can alleviate the interference of noise relation annotations via label softening. Nevertheless, we find that they still suffer from two problems: propagation of inaccurate dark knowledge and constraint of a unified distillation temperature. In this article, we propose a simple and effective Multi-instance Dynamic Temperature Distillation (MiDTD) framework, which is model-agnostic and mainly involves two modules: multi-instance target fusion (MiTF) and dynamic temperature regulation (DTR). MiTF combines the teacher's predictions for multiple sentences with the same entity pair to amend the inaccurate dark knowledge in each student's target. DTR allocates alterable distillation temperatures to different training instances to enable the softness of most student's targets to be regulated to a moderate range. In experiments, we construct three concrete MiDTD instantiations with BERT, PCNN, and BiLSTM-based RE models, and the distilled students significantly outperform their teachers and the state-of-the-art (SOTA) methods.
引用
收藏
页数:32
相关论文
共 50 条
  • [41] Using Dilated Residual Network to Model Distantly Supervised Relation Extraction
    Zhan, Lei
    Yang, Yan
    Zhu, Pinpin
    He, Liang
    Yu, Zhou
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, 2019, 11448 : 500 - 504
  • [42] Active Testing: An Unbiased Evaluation Method for Distantly Supervised Relation Extraction
    Li, Pengshuai
    Zhang, Xinsong
    Jia, Weijia
    Zhao, Wei
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 204 - 211
  • [43] DiS-ReX: A Multilingual Dataset for Distantly Supervised Relation Extraction
    Bhartiya, Abhyuday
    Badola, Kartikeya
    Mausam
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022): (SHORT PAPERS), VOL 2, 2022, : 849 - 863
  • [44] Knowing False Negatives: An Adversarial Training Method for Distantly Supervised Relation Extraction
    Hao, Kailong
    Yu, Botao
    Hu, Wei
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 9661 - 9672
  • [45] Improving Distantly Supervised Relation Extraction with Multi-Level Noise Reduction
    Song, Wei
    Yang, Zijiang
    AI, 2024, 5 (03) : 1709 - 1730
  • [46] Learning with Noise: Enhance Distantly Supervised Relation Extraction with Dynamic Transition Matrix
    Luo, Bingfeng
    Feng, Yansong
    Wang, Zheng
    Zhu, Zhanxing
    Huang, Songfang
    Yan, Rui
    Zhao, Dongyan
    PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 1, 2017, : 430 - 439
  • [47] Cross-Relation Cross-Bag Attention for Distantly-Supervised Relation Extraction
    Yuan, Yujin
    Liu, Liyuan
    Tang, Siliang
    Zhang, Zhongfei
    Zhuang, Yueting
    Pu, Shiliang
    Wu, Fei
    Ren, Xiang
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 419 - 426
  • [48] Manual Evaluation Matters: Reviewing Test Protocols of Distantly Supervised Relation Extraction
    Gao, Tianyu
    Han, Xu
    Bai, Yuzhuo
    Qiu, Keyue
    Xie, Zhiyu
    Lin, Yankai
    Liu, Zhiyuan
    Li, Peng
    Sun, Maosong
    Zhou, Jie
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 1306 - 1318
  • [49] Large-scale Opinion Relation Extraction with Distantly Supervised Neural Network
    Sun, Changzhi
    Wu, Yuanbin
    Lan, Man
    Sun, Shiliang
    Zhang, Qi
    15TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2017), VOL 1: LONG PAPERS, 2017, : 1033 - 1043
  • [50] Distantly Supervised relation extraction with multi-level contextual information integration
    Han, Danjie
    Huang, Heyan
    Shi, Shumin
    Yuan, Changsen
    Guo, Cunhan
    NEUROCOMPUTING, 2025, 634