A pattern-aware self-attention network for distant supervised relation extraction

被引:30
|
作者
Shang, Yu-Ming [1 ]
Huang, Heyan [1 ,2 ]
Sun, Xin [1 ]
Wei, Wei [3 ]
Mao, Xian-Ling [1 ]
机构
[1] Beijing Inst Technol, Sch Comp Sci, Beijing, Peoples R China
[2] Beijing Engn Res Ctr High Volume Language Informa, Beijing, Peoples R China
[3] Huazhong Univ Sci & Technol, Huazhong, Hubei, Peoples R China
关键词
distant supervision; relation extraction; pre-trained Transformer; relational pattern; self-attention network;
D O I
10.1016/j.ins.2021.10.047
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Distant supervised relation extraction is an efficient strategy of finding relational facts from unstructured text without labeled training data. A recent paradigm to develop relation extractors is using pre-trained Transformer language models to produce high-quality sentence representations. However, due to the original Transformer is weak at capturing local dependencies and phrasal structures, existing Transformer-based methods cannot identify various relational patterns in sentences. To address this issue, we propose a novel distant supervised relation extraction model, which employs a specific-designed pattern-aware self-attention network to automatically discover relational patterns for pre-trained Transformers in an end-to-end manner. Specifically, the proposed method assumes that the correlation between two adjacent tokens reflects the probability that they belong to the same pattern. Based on this assumption, a novel self-attention network is designed to generate the probability distribution of all patterns in a sentence. Then, the probability distribution is applied as a constraint in the first Transformer layer to encourage its attention heads to follow the relational pattern structures. As a result, fine-grained pattern information is enhanced in the pre-trained Transformer without losing global dependencies. Extensive experimental results on two popular benchmark datasets demonstrate that our model performs better than the state-of-the-art baselines. (c) 2021 Elsevier Inc. All rights reserved.
引用
收藏
页码:269 / 279
页数:11
相关论文
共 50 条
  • [31] Multi-Path Convolutional Neural Network for Distant Supervised Relation Extraction
    Li, Yunyang
    Zhong, Zhinong
    Jing, Ning
    PROCEEDINGS OF THE 2ND INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND APPLICATION ENGINEERING (CSAE2018), 2018,
  • [32] Pattern-Aware Reliable Virtual Network Function Chain Deployment
    Zhang, Xin
    Qian, Zhuzhong
    Zhang, Sheng
    Lu, Sanglu
    2017 15TH IEEE INTERNATIONAL SYMPOSIUM ON PARALLEL AND DISTRIBUTED PROCESSING WITH APPLICATIONS AND 2017 16TH IEEE INTERNATIONAL CONFERENCE ON UBIQUITOUS COMPUTING AND COMMUNICATIONS (ISPA/IUCC 2017), 2017, : 335 - 342
  • [33] Context-Aware Self-Attention Networks
    Yang, Baosong
    Li, Jian
    Wong, Derek F.
    Chao, Lidia S.
    Wang, Xing
    Tu, Zhaopeng
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 387 - 394
  • [34] Distantly Supervised Relation Extraction Based on Residual Attention and Self Learning
    Zheng, Zhiyun
    Xu, Yamei
    Liu, Yun
    Zhang, Xingjin
    Li, Lun
    Li, Dun
    NEURAL PROCESSING LETTERS, 2024, 56 (03)
  • [35] Understanding Self-Attention of Self-Supervised Audio Transformers
    Yang, Shu-wen
    Liu, Andy T.
    Lee, Hung-yi
    INTERSPEECH 2020, 2020, : 3785 - 3789
  • [36] A Supervised Multi-Head Self-Attention Network for Nested Named Entity Recognition
    Xu, Yongxiu
    Huang, Heyan
    Feng, Chong
    Hu, Yue
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 14185 - 14193
  • [37] Saliency Guided Self-Attention Network for Weakly and Semi-Supervised Semantic Segmentation
    Yao, Qi
    Gong, Xiaojin
    IEEE ACCESS, 2020, 8 : 14413 - 14423
  • [38] Relation Extraction using Multi-Encoder LSTM Network on a Distant Supervised Dataset
    Banerjee, Siddhartha
    Tsioutsiouliklis, Kostas
    2018 IEEE 12TH INTERNATIONAL CONFERENCE ON SEMANTIC COMPUTING (ICSC), 2018, : 235 - 238
  • [39] Weakly supervised histopathology image segmentation with self-attention
    Li, Kailu
    Qian, Ziniu
    Han, Yingnan
    Chang, Eric I-Chao
    Wei, Bingzheng
    Lai, Maode
    Liao, Jing
    Fan, Yubo
    Xu, Yan
    MEDICAL IMAGE ANALYSIS, 2023, 86
  • [40] Self-attention Hypergraph Pooling Network
    Zhao Y.-F.
    Jin F.-S.
    Li R.-H.
    Qin H.-C.
    Cui P.
    Wang G.-R.
    Ruan Jian Xue Bao/Journal of Software, 2023, 34 (10):