Pre-training a Neural Model to Overcome Data Scarcity in Relation Extraction from Text

被引:0
|
作者
Jung, Seokwoo [1 ,2 ]
Myaeng, Sung-Hyon [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Sch Comp, Daejeon, South Korea
[2] NAVER R&D Ctr, AiRS Ai Recommender Syst, Seongnam, South Korea
基金
新加坡国家研究基金会;
关键词
relation extraction; pre-training; unsupervised earning; dependency parsing; sentence embedding; pcnn;
D O I
10.1109/bigcomp.2019.8679242
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Data scarcity is a major stumbling block in relation extraction. We propose an unsupervised pre-training method for extracting relational information from a huge amount of unlabeled data prior to supervised learning in the situation where hard to make golden labeled data. An objective function not requiring any labeled data is adopted during the pre-training phase, with an attempt to predict clue words crucial for inferring semantic relation types between two entities in a given sentence. The experimental result on public datasets shows that our approach achieves similar performance by using only 70% of data in a data-scarce setting.
引用
下载
收藏
页码:176 / 180
页数:5
相关论文
共 50 条
  • [1] A Method of Relation Extraction Using Pre-training Models
    Wang, Yu
    Sun, Yining
    Ma, Zuchang
    Gao, Lisheng
    Xu, Yang
    Wu, Yichen
    2020 13TH INTERNATIONAL SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE AND DESIGN (ISCID 2020), 2020, : 176 - 179
  • [2] A STUDY ON THE EFFICACY OF MODEL PRE-TRAINING IN DEVELOPING NEURAL TEXT-TO-SPEECH SYSTEM
    Zhang, Guangyan
    Leng, Yichong
    Tan, Daxin
    Qin, Ying
    Song, Kaitao
    Tan, Xu
    Zhao, Sheng
    Lee, Tan
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 6087 - 6091
  • [3] Relation Extraction with Weighted Contrastive Pre-training on Distant Supervision
    Wan, Zhen
    Cheng, Fei
    Liu, Qianying
    Mao, Zhuoyuan
    Song, Haiyue
    Kurohashi, Sadao
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 2580 - 2585
  • [4] Unifying Structured Data as Graph for Data-to-Text Pre-Training
    Li, Shujie
    Li, Liang
    Geng, Ruiying
    Yang, Min
    Li, Binhua
    Yuan, Guanghu
    He, Wanwei
    Yuan, Shao
    Ma, Can
    Huang, Fei
    Li, Yongbin
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2024, 12 : 210 - 228
  • [5] Investigation of improving the pre-training and fine-tuning of BERT model for biomedical relation extraction
    Peng Su
    K. Vijay-Shanker
    BMC Bioinformatics, 23
  • [6] Relational distance and document-level contrastive pre-training based relation extraction model
    Dong, Yihao
    Xu, Xiaolong
    PATTERN RECOGNITION LETTERS, 2023, 167 : 132 - 140
  • [7] Investigation of improving the pre-training and fine-tuning of BERT model for biomedical relation extraction
    Su, Peng
    Vijay-Shanker, K.
    BMC BIOINFORMATICS, 2022, 23 (01)
  • [8] Pre-training to Match for Unified Low-shot Relation Extraction
    Liu, Fangchao
    Lin, Hongyu
    Han, Xianpei
    Cao, Boxi
    Sun, Le
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 5785 - 5795
  • [9] Numerical Tuple Extraction from Tables with Pre-training
    Yang, Qingping
    Cao, Yixuan
    Luo, Ping
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 2233 - 2241
  • [10] PRE-TRAINING TRANSFORMER DECODER FOR END-TO-END ASR MODEL WITH UNPAIRED TEXT DATA
    Gao, Changfeng
    Cheng, Gaofeng
    Yang, Runyan
    Zhu, Han
    Zhang, Pengyuan
    Yan, Yonghong
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 6543 - 6547