A Positive-Unlabeled Metric Learning Framework for Document-Level Relation Extraction with Incomplete Labeling

被引:0
|
作者
Wang, Ye [1 ]
Pan, Huazheng [1 ]
Zhang, Tao [2 ]
Wu, Wen [1 ]
Hu, Wenxin [1 ]
机构
[1] East China Normal Univ, Shanghai, Peoples R China
[2] Tsinghua Univ, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The goal of document-level relation extraction (RE) is to identify relations between entities that span multiple sentences. Recently, incomplete labeling in document-level RE has received increasing attention, and some studies have used methods such as positive-unlabeled learning to tackle this issue, but there is still a lot of room for improvement. Motivated by this, we propose a positive-augmentation and positive-mixup positive-unlabeled metric learning framework ((PM)-M-3). Specifically, we formulate document-level RE as a metric learning problem. We aim to pull the distance closer between entity pair embedding and their corresponding relation embedding, while pushing it farther away from the none class relation embedding. Additionally, we adapt the positive unlabeled learning to this loss objective. In order to improve the generalizability of the model, we use dropout to augment positive samples and propose a positive-none-class mixup method. Extensive experiments show that P3M improves the Fl score by approximately 4-10 points in document-level RE with incomplete labeling, and achieves state-of-the-art results in fully labeled scenarios. Furthermore, P3M has also demonstrated robustness to prior estimation bias in incomplete labeled scenarios.
引用
收藏
页码:19197 / 19205
页数:9
相关论文
共 50 条
  • [31] Evidence-aware Document-level Relation Extraction
    Xu, Tianyu
    Hua, Wen
    Qu, Jianfeng
    Li, Zhixu
    Xu, Jiajie
    Liu, An
    Zhao, Lei
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 2311 - 2320
  • [32] HistRED: A Historical Document-Level Relation Extraction Dataset
    Yang, Soyoung
    Choi, Minseok
    Cho, Youngwoo
    Choo, Jaegul
    arXiv, 2023,
  • [33] A Hierarchical Network for Multimodal Document-Level Relation Extraction
    Kong, Lingxing
    Wang, Jiuliang
    Ma, Zheng
    Zhou, Qifeng
    Zhang, Jianbing
    He, Liang
    Chen, Jiajun
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 16, 2024, : 18408 - 18416
  • [34] HistRED: A Historical Document-Level Relation Extraction Dataset
    Yang, Soyoung
    Choi, Minseok
    Cho, Youngwoo
    Choo, Jaegul
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 3207 - 3224
  • [35] Rethinking Document-Level Relation Extraction: A Reality Check
    Li, Jing
    Wang, Yequan
    Zhang, Shuai
    Zhang, Min
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, 2023, : 5715 - 5730
  • [36] An adaptive confidence-based data revision framework for Document-level Relation Extraction
    Jiang, Chao
    Liao, Jinzhi
    Zhao, Xiang
    Zeng, Daojian
    Dai, Jianhua
    INFORMATION PROCESSING & MANAGEMENT, 2025, 62 (01)
  • [37] Collective prompt tuning with relation inference for document-level relation extraction
    Yuan, Changsen
    Cao, Yixin
    Huang, Heyan
    INFORMATION PROCESSING & MANAGEMENT, 2023, 60 (05)
  • [38] Multi-View Cooperative Learning with Invariant Rationale for Document-Level Relation Extraction
    Lin, Rui
    Fan, Jing
    He, Yinglong
    Yang, Yehui
    Li, Jia
    Guo, Cunhan
    COGNITIVE COMPUTATION, 2024, 16 (06) : 3505 - 3517
  • [39] Principled analytic classifier for positive-unlabeled learning via weighted integral probability metric
    Yongchan Kwon
    Wonyoung Kim
    Masashi Sugiyama
    Myunghee Cho Paik
    Machine Learning, 2020, 109 : 513 - 532
  • [40] End-to-end Learning of Logical Rules for Enhancing Document-level Relation Extraction
    Qi, Kunxun
    Du, Jianfeng
    Wan, Hai
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1: LONG PAPERS, 2024, : 7247 - 7263