MKDAT: Multi-Level Knowledge Distillation with Adaptive Temperature for Distantly Supervised Relation Extraction

被引:0
|
作者
Long, Jun [1 ]
Yin, Zhuoying [1 ,2 ]
Han, Yan [3 ]
Huang, Wenti [4 ]
机构
[1] Cent South Univ, Big Data Inst, Changsha 410075, Peoples R China
[2] Guizhou Rural Credit Union, Guiyang 550000, Peoples R China
[3] Guizhou Univ Commerce, Sch Comp & Informat Engn, Guiyang 550025, Peoples R China
[4] Hunan Univ Sci & Technol, Sch Comp Sci & Engn, Xiangtan 411100, Peoples R China
基金
中国国家自然科学基金;
关键词
distantly supervised relation extraction; knowledge distillation; label softening;
D O I
10.3390/info15070382
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Distantly supervised relation extraction (DSRE), first used to address the limitations of manually annotated data via automatically annotating the data with triplet facts, is prone to issues such as mislabeled annotations due to the interference of noisy annotations. To address the interference of noisy annotations, we leveraged a novel knowledge distillation (KD) method which was different from the conventional models on DSRE. More specifically, we proposed a model-agnostic KD method, Multi-Level Knowledge Distillation with Adaptive Temperature (MKDAT), which mainly involves two modules: Adaptive Temperature Regulation (ATR) and Multi-Level Knowledge Distilling (MKD). ATR allocates adaptive entropy-based distillation temperatures to different training instances for providing a moderate softening supervision to the student, in which label hardening is possible for instances with great entropy. MKD combines the bag-level and instance-level knowledge of the teacher as supervisions of the student, and trains the teacher and student at the bag and instance levels, respectively, which aims at mitigating the effects of noisy annotation and improving the sentence-level prediction performance. In addition, we implemented three MKDAT models based on the CNN, PCNN, and ATT-BiLSTM neural networks, respectively, and the experimental results show that our distillation models outperform the baseline models on bag-level and instance-level evaluations.
引用
收藏
页数:18
相关论文
共 50 条
  • [31] Semantic piecewise convolutional neural network with adaptive negative training for distantly supervised relation extraction
    Yu, Mei
    Chen, Yunke
    Zhao, Mankun
    Xu, Tianyi
    Yu, Jian
    Yu, Ruiguo
    Liu, Hongwei
    Li, Xuewei
    NEUROCOMPUTING, 2023, 537 : 12 - 21
  • [32] Multiresolution Knowledge Distillation and Multi-level Fusion for Defect Detection
    Xie, Huosheng
    Xiao, Yan
    GREEN, PERVASIVE, AND CLOUD COMPUTING, GPC 2022, 2023, 13744 : 178 - 191
  • [33] Populating Web-Scale Knowledge Graphs Using Distantly Supervised Relation Extraction and Validation
    Dash, Sarthak
    Glass, Michael R.
    Gliozzo, Alfio
    Canim, Mustafa
    Rossiello, Gaetano
    INFORMATION, 2021, 12 (08)
  • [34] Local-to-global GCN with knowledge-aware representation for distantly supervised relation extraction
    Huang, Wenti
    Mao, Yiyu
    Yang, Liu
    Yang, Zhan
    Long, Jun
    KNOWLEDGE-BASED SYSTEMS, 2021, 234
  • [35] Improving Supervised Drug-Protein Relation Extraction with Distantly Supervised Models
    Iinuma, Naoki
    Miwa, Makoto
    Sasaki, Yutaka
    PROCEEDINGS OF THE 21ST WORKSHOP ON BIOMEDICAL LANGUAGE PROCESSING (BIONLP 2022), 2022, : 161 - 170
  • [36] Distantly Supervised Entity Relation Extraction with Adapted Manual Annotations
    Sun, Changzhi
    Wu, Yuanbin
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 7039 - 7046
  • [37] Denoising by Markov Random Filed in Distantly Supervised Relation Extraction
    Li, Yameng
    Liu, Ruifang
    PROCEEDINGS OF 2017 3RD IEEE INTERNATIONAL CONFERENCE ON COMPUTER AND COMMUNICATIONS (ICCC), 2017, : 1525 - 1529
  • [38] Exploring Long Tail Data in Distantly Supervised Relation Extraction
    Gui, Yaocheng
    Liu, Qian
    Zhu, Man
    Gao, Zhiqiang
    NATURAL LANGUAGE UNDERSTANDING AND INTELLIGENT APPLICATIONS (NLPCC 2016), 2016, 10102 : 514 - 522
  • [39] MSnet: Multi-Head Self-Attention Network for Distantly Supervised Relation Extraction
    Sun, Tingting
    Zhang, Chunhong
    Ji, Yang
    Hu, Zheng
    IEEE ACCESS, 2019, 7 : 54472 - 54482
  • [40] Improving Distantly Supervised Relation Extraction by Natural Language Inference
    Zhou, Kang
    Qiao, Qiao
    Li, Yuepei
    Li, Qi
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 11, 2023, : 14047 - 14055