MKDAT: Multi-Level Knowledge Distillation with Adaptive Temperature for Distantly Supervised Relation Extraction

被引:0
|
作者
Long, Jun [1 ]
Yin, Zhuoying [1 ,2 ]
Han, Yan [3 ]
Huang, Wenti [4 ]
机构
[1] Cent South Univ, Big Data Inst, Changsha 410075, Peoples R China
[2] Guizhou Rural Credit Union, Guiyang 550000, Peoples R China
[3] Guizhou Univ Commerce, Sch Comp & Informat Engn, Guiyang 550025, Peoples R China
[4] Hunan Univ Sci & Technol, Sch Comp Sci & Engn, Xiangtan 411100, Peoples R China
基金
中国国家自然科学基金;
关键词
distantly supervised relation extraction; knowledge distillation; label softening;
D O I
10.3390/info15070382
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Distantly supervised relation extraction (DSRE), first used to address the limitations of manually annotated data via automatically annotating the data with triplet facts, is prone to issues such as mislabeled annotations due to the interference of noisy annotations. To address the interference of noisy annotations, we leveraged a novel knowledge distillation (KD) method which was different from the conventional models on DSRE. More specifically, we proposed a model-agnostic KD method, Multi-Level Knowledge Distillation with Adaptive Temperature (MKDAT), which mainly involves two modules: Adaptive Temperature Regulation (ATR) and Multi-Level Knowledge Distilling (MKD). ATR allocates adaptive entropy-based distillation temperatures to different training instances for providing a moderate softening supervision to the student, in which label hardening is possible for instances with great entropy. MKD combines the bag-level and instance-level knowledge of the teacher as supervisions of the student, and trains the teacher and student at the bag and instance levels, respectively, which aims at mitigating the effects of noisy annotation and improving the sentence-level prediction performance. In addition, we implemented three MKDAT models based on the CNN, PCNN, and ATT-BiLSTM neural networks, respectively, and the experimental results show that our distillation models outperform the baseline models on bag-level and instance-level evaluations.
引用
收藏
页数:18
相关论文
共 50 条
  • [21] Multi-Level Knowledge Distillation with Positional Encoding Enhancement
    Xu, Lixiang
    Wang, Zhiwen
    Bai, Lu
    Ji, Shengwei
    Ai, Bing
    Wang, Xiaofeng
    Yu, Philip S.
    PATTERN RECOGNITION, 2025, 163
  • [22] Knowledge-Aware and Retrieval-Based Models for Distantly Supervised Relation Extraction
    Zhang, Xuemiao
    Deng, Kejun
    Zhang, Leilei
    Tan, Zhouxing
    Liu, Junfei
    PRICAI 2019: TRENDS IN ARTIFICIAL INTELLIGENCE, PT I, 2019, 11670 : 148 - 161
  • [23] Revisiting the Negative Data of Distantly Supervised Relation Extraction
    Xie, Chenhao
    Liang, Jiaqing
    Liu, Jingping
    Huang, Chengsong
    Huang, Wenhao
    Xiao, Yanghua
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 3572 - 3581
  • [24] Learning labeling functions in distantly supervised relation extraction
    Gui, Yaocheng
    Liu, Qian
    Gao, Zhiqiang
    INTELLIGENT DATA ANALYSIS, 2020, 24 (02) : 427 - 443
  • [25] Incorporating Instance Correlations in Distantly Supervised Relation Extraction
    Zhang, Luhao
    Hu, Linmei
    Shi, Chuan
    SEMANTIC TECHNOLOGY, JIST 2019: PROCEEDINGS, 2020, 12032 : 177 - 191
  • [26] Dynamic Chunkwise CNN for Distantly Supervised Relation Extraction
    Liu, Fangbing
    Wang, Qing
    2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2020, : 738 - 747
  • [27] Type affinity network for distantly supervised relation extraction
    Song, Wei
    Zhou, Juncheng
    Liu, Xiao
    NEUROCOMPUTING, 2025, 630
  • [28] Distantly Supervised Neural Network Model for Relation Extraction
    Wang, Zhen
    Chang, Baobao
    Sui, Zhifang
    CHINESE COMPUTATIONAL LINGUISTICS AND NATURAL LANGUAGE PROCESSING BASED ON NATURALLY ANNOTATED BIG DATA (CCL 2015), 2015, 9427 : 253 - 266
  • [29] A Multi-Level Adaptive Lightweight Net for Damaged Road Marking Detection Based on Knowledge Distillation
    Wang, Junwei
    Zeng, Xiangqiang
    Wang, Yong
    Ren, Xiang
    Wang, Dongliang
    Qu, Wenqiu
    Liao, Xiaohan
    Pan, Peifen
    REMOTE SENSING, 2024, 16 (14)
  • [30] Improving Distantly Supervised Relation Extraction by Knowledge Base-Driven Zero Subject Resolution
    Kim, Eun-Kyung
    Choi, Key-Sun
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2018, E101D (10) : 2551 - 2558