MKDAT: Multi-Level Knowledge Distillation with Adaptive Temperature for Distantly Supervised Relation Extraction

被引:0
|
作者
Long, Jun [1 ]
Yin, Zhuoying [1 ,2 ]
Han, Yan [3 ]
Huang, Wenti [4 ]
机构
[1] Cent South Univ, Big Data Inst, Changsha 410075, Peoples R China
[2] Guizhou Rural Credit Union, Guiyang 550000, Peoples R China
[3] Guizhou Univ Commerce, Sch Comp & Informat Engn, Guiyang 550025, Peoples R China
[4] Hunan Univ Sci & Technol, Sch Comp Sci & Engn, Xiangtan 411100, Peoples R China
基金
中国国家自然科学基金;
关键词
distantly supervised relation extraction; knowledge distillation; label softening;
D O I
10.3390/info15070382
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Distantly supervised relation extraction (DSRE), first used to address the limitations of manually annotated data via automatically annotating the data with triplet facts, is prone to issues such as mislabeled annotations due to the interference of noisy annotations. To address the interference of noisy annotations, we leveraged a novel knowledge distillation (KD) method which was different from the conventional models on DSRE. More specifically, we proposed a model-agnostic KD method, Multi-Level Knowledge Distillation with Adaptive Temperature (MKDAT), which mainly involves two modules: Adaptive Temperature Regulation (ATR) and Multi-Level Knowledge Distilling (MKD). ATR allocates adaptive entropy-based distillation temperatures to different training instances for providing a moderate softening supervision to the student, in which label hardening is possible for instances with great entropy. MKD combines the bag-level and instance-level knowledge of the teacher as supervisions of the student, and trains the teacher and student at the bag and instance levels, respectively, which aims at mitigating the effects of noisy annotation and improving the sentence-level prediction performance. In addition, we implemented three MKDAT models based on the CNN, PCNN, and ATT-BiLSTM neural networks, respectively, and the experimental results show that our distillation models outperform the baseline models on bag-level and instance-level evaluations.
引用
收藏
页数:18
相关论文
共 50 条
  • [41] ReadsRE: Retrieval-Augmented Distantly Supervised Relation Extraction
    Zhang, Yue
    Fei, Hongliang
    Li, Ping
    SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, : 2257 - 2262
  • [42] Interaction-and-Response Network for Distantly Supervised Relation Extraction
    Song, Wei
    Gu, Weishuai
    Zhu, Fuxin
    Park, Soon Cheol
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (07) : 9523 - 9537
  • [43] Empower Distantly Supervised Relation Extraction with Collaborative Adversarial Training
    Chen, Tao
    Shi, Haochen
    Liu, Liyuan
    Tang, Siliang
    Shao, Jian
    Chen, Zhigang
    Zhuang, Yueting
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 12675 - 12682
  • [44] Multi-level Logit Distillation
    Jin, Ying
    Wang, Jiaqi
    Lin, Dahua
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 24276 - 24285
  • [45] Bootstrapped Multi-level Distant Supervision for Relation Extraction
    He, Ying
    Li, Zhixu
    Liu, Guanfeng
    Cao, Fangfei
    Chen, Zhigang
    Wang, Ke
    Ma, Jie
    WEB INFORMATION SYSTEMS ENGINEERING, WISE 2018, PT I, 2018, 11233 : 408 - 423
  • [46] Distantly Supervised Relation Extraction via Contextual Information Interaction and Relation Embeddings
    Yin, Huixin
    Liu, Shengquan
    Jian, Zhaorui
    SYMMETRY-BASEL, 2023, 15 (09):
  • [47] Multi-Level Knowledge Distillation for Speech Emotion Recognition in Noisy Conditions
    Liu, Yang
    Sun, Haoqin
    Chen, Geng
    Wang, Qingyue
    Zhao, Zhen
    Lu, Xugang
    Wang, Longbiao
    INTERSPEECH 2023, 2023, : 1893 - 1897
  • [48] Multi-Level Knowledge Distillation for Out-of-Distribution Detection in Text
    Wu, Qianhui
    Jiang, Huiqiang
    Yin, Haonan
    Karlsson, Borje F.
    Lin, Chin-Yew
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 7317 - 7332
  • [49] Adversarial Multi-Teacher Distillation for Semi-Supervised Relation Extraction
    Li, Wanli
    Qian, Tieyun
    Li, Xuhui
    Zou, Lixin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (08) : 11291 - 11301
  • [50] Noise-Robust Semi-Supervised Learning for Distantly Supervised Relation Extraction
    Sun, Xin
    Liu, Qiang
    Wu, Shu
    Wang, Zilei
    Wang, Liang
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 13145 - 13157