A comprehensive exploration of semantic relation extraction via pre-trained CNNs

被引:32
|
作者
Li, Qing [1 ]
Li, Lili [2 ]
Wang, Weinan [3 ]
Li, Qi [4 ]
Zhong, Jiang [1 ,5 ]
机构
[1] Chongqing Univ, Coll Comp Sci, Chongqing, Peoples R China
[2] Chongqing Univ, Sch Civil Engn, Chongqing, Peoples R China
[3] Peking Univ, Sch Math Sci, Beijing, Peoples R China
[4] Shaoxing Univ, Dept Comp Sci & Engn, Shaoxing, Peoples R China
[5] Chongqing Univ, Key Lab Dependable Serv Comp Cyber Phys Soc, Chongqing, Peoples R China
关键词
Relation extraction; Semantic relation; Natural language processing; Convolutional neural networks;
D O I
10.1016/j.knosys.2020.105488
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Semantic relation extraction between entity pairs is a crucial task in information extraction from text. In this paper, we propose a new pre-trained network architecture for this task, and it is called the XM-CNN. The XM-CNN utilizes word embedding and position embedding information. It is designed to reinforce the contextual output from the MT-DNNKD pre-trained model. Our model effectively utilized an entity-aware attention mechanisms to detected the features and also adopts and applies more relation-specific pooling attention mechanisms applied to it. The experimental results show that the XM-CNN achieves state-of-the-art results on the SemEval-2010 task 8, and a thorough evaluation of the method is conducted. (C) 2020 Elsevier B.V. All rights reserved.
引用
收藏
页数:7
相关论文
共 50 条
  • [41] Implementation of CNNs for Crop Diseases Classification: A Comparison of Pre-trained Model and Training from Scratch
    Sahu, Priyanka
    Chug, Anuradha
    Singh, Amit Prakash
    Singh, Dinesh
    Singh, Ravinder Pal
    INTERNATIONAL JOURNAL OF COMPUTER SCIENCE AND NETWORK SECURITY, 2020, 20 (10): : 206 - 215
  • [42] Evaluation of transfer learning of pre-trained CNNs applied to breast cancer detection on infrared images
    Chaves, Esdras
    Goncalves, Caroline B.
    Albertini, Marcelo K.
    Lee, Soojeong
    Jeon, Gwanggil
    Fernandes, Henrique C.
    APPLIED OPTICS, 2020, 59 (17) : E23 - E28
  • [43] Surface-aware Mesh Texture Synthesis with Pre-trained 2D CNNs
    Kovacs, Aron Samuel
    Hermosilla, Pedro
    Raidou, Renata G.
    COMPUTER GRAPHICS FORUM, 2024, 43 (02)
  • [44] Multilingual Translation via Grafting Pre-trained Language Models
    Sun, Zewei
    Wang, Mingxuan
    Li, Lei
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 2735 - 2747
  • [45] Compression of Generative Pre-trained Language Models via Quantization
    Tao, Chaofan
    Hou, Lu
    Zhang, Wei
    Shang, Lifeng
    Jiang, Xin
    Liu, Qun
    Luo, Ping
    Wong, Ngai
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 4821 - 4836
  • [46] Parallel Corpus Filtering via Pre-trained Language Models
    DiDi Labs
    arXiv, 2020,
  • [47] Trend and seasonality features extraction with pre-trained CNN and recurrence plot
    Strozzi, Fernanda
    Pozzi, Rossella
    INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2024, 62 (09) : 3251 - 3262
  • [48] ASIC IMPLEMENTATION OF A PRE-TRAINED NEURAL NETWORK FOR ECG FEATURE EXTRACTION
    Tefai, Huruy Tekle
    Saleh, Hani
    Tekeste, Temesghen
    Alqutayri, Mahmoud
    Mohammad, Baker
    2020 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2020,
  • [49] Semantic Importance-Aware Communications Using Pre-Trained Language Models
    Guo, Shuaishuai
    Wang, Yanhu
    Li, Shujing
    Saeed, Nasir
    IEEE COMMUNICATIONS LETTERS, 2023, 27 (09) : 2328 - 2332
  • [50] Enriching Pre-trained Language Model with Entity Information for Relation Classification
    Wu, Shanchan
    He, Yifan
    PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT (CIKM '19), 2019, : 2361 - 2364