Pre-training a Neural Model to Overcome Data Scarcity in Relation Extraction from Text

被引:0
|
作者
Jung, Seokwoo [1 ,2 ]
Myaeng, Sung-Hyon [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Sch Comp, Daejeon, South Korea
[2] NAVER R&D Ctr, AiRS Ai Recommender Syst, Seongnam, South Korea
基金
新加坡国家研究基金会;
关键词
relation extraction; pre-training; unsupervised earning; dependency parsing; sentence embedding; pcnn;
D O I
10.1109/bigcomp.2019.8679242
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Data scarcity is a major stumbling block in relation extraction. We propose an unsupervised pre-training method for extracting relational information from a huge amount of unlabeled data prior to supervised learning in the situation where hard to make golden labeled data. An objective function not requiring any labeled data is adopted during the pre-training phase, with an attempt to predict clue words crucial for inferring semantic relation types between two entities in a given sentence. The experimental result on public datasets shows that our approach achieves similar performance by using only 70% of data in a data-scarce setting.
引用
下载
收藏
页码:176 / 180
页数:5
相关论文
共 50 条
  • [31] Pre-Training on Mixed Data for Low-Resource Neural Machine Translation
    Zhang, Wenbo
    Li, Xiao
    Yang, Yating
    Dong, Rui
    INFORMATION, 2021, 12 (03)
  • [32] A deep neural network model for Chinese toponym matching with geographic pre-training model
    Qiu, Qinjun
    Zheng, Shiyu
    Tian, Miao
    Li, Jiali
    Ma, Kai
    Tao, Liufeng
    Xie, Zhong
    INTERNATIONAL JOURNAL OF DIGITAL EARTH, 2024, 17 (01)
  • [33] Multiple Document Datasets Pre-training Improves Text Line Detection With Deep Neural Networks
    Boillet, Melodie
    Kermorvant, Christopher
    Paquet, Thierry
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 2134 - 2141
  • [34] Research on Effectiveness of the Pre-Training Model in Improving the Performance of Spectral Feature Extraction
    Ren, Ju-Xiang
    Liu, Zhong-Bao
    Guang Pu Xue Yu Guang Pu Fen Xi/Spectroscopy and Spectral Analysis, 2024, 44 (12): : 3480 - 3484
  • [35] Neural speech enhancement with unsupervised pre-training and mixture training
    Hao, Xiang
    Xu, Chenglin
    Xie, Lei
    NEURAL NETWORKS, 2023, 158 : 216 - 227
  • [36] Surgery of Speech Synthesis Models to Overcome the Scarcity of Training Data
    Pierard, Arnaud
    Erro, D.
    Hernaez, I.
    Navas, E.
    Dutoit, Thierry
    ADVANCES IN SPEECH AND LANGUAGE TECHNOLOGIES FOR IBERIAN LANGUAGES, IBERSPEECH 2016, 2016, 10077 : 73 - 83
  • [37] A Multi-Channel Text Sentiment Analysis Model Integrating Pre-training Mechanism
    Liang, Shengbin
    Jin, Jiangyong
    Du, Wencai
    Qu, Shenming
    INFORMATION TECHNOLOGY AND CONTROL, 2023, 52 (02): : 263 - 275
  • [38] Medical text classification based on the discriminative pre-training model and prompt-tuning
    Wang, Yu
    Wang, Yuan
    Peng, Zhenwan
    Zhang, Feifan
    Zhou, Luyao
    Yang, Fei
    DIGITAL HEALTH, 2023, 9
  • [39] ASR-Generated Text for Language Model Pre-training Applied to Speech Tasks
    Pelloin, Valentin
    Dary, Franck
    Herve, Nicolas
    Favre, Benoit
    Camelin, Nathalie
    Laurent, Antoine
    Besacier, Laurent
    INTERSPEECH 2022, 2022, : 3453 - 3457
  • [40] Virtual prompt pre-training for prototype-based few-shot relation extraction
    He, Kai
    Huang, Yucheng
    Mao, Rui
    Gong, Tieliang
    Li, Chen
    Cambria, Erik
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 213