A pattern-aware self-attention network for distant supervised relation extraction

被引:30
|
作者
Shang, Yu-Ming [1 ]
Huang, Heyan [1 ,2 ]
Sun, Xin [1 ]
Wei, Wei [3 ]
Mao, Xian-Ling [1 ]
机构
[1] Beijing Inst Technol, Sch Comp Sci, Beijing, Peoples R China
[2] Beijing Engn Res Ctr High Volume Language Informa, Beijing, Peoples R China
[3] Huazhong Univ Sci & Technol, Huazhong, Hubei, Peoples R China
关键词
distant supervision; relation extraction; pre-trained Transformer; relational pattern; self-attention network;
D O I
10.1016/j.ins.2021.10.047
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Distant supervised relation extraction is an efficient strategy of finding relational facts from unstructured text without labeled training data. A recent paradigm to develop relation extractors is using pre-trained Transformer language models to produce high-quality sentence representations. However, due to the original Transformer is weak at capturing local dependencies and phrasal structures, existing Transformer-based methods cannot identify various relational patterns in sentences. To address this issue, we propose a novel distant supervised relation extraction model, which employs a specific-designed pattern-aware self-attention network to automatically discover relational patterns for pre-trained Transformers in an end-to-end manner. Specifically, the proposed method assumes that the correlation between two adjacent tokens reflects the probability that they belong to the same pattern. Based on this assumption, a novel self-attention network is designed to generate the probability distribution of all patterns in a sentence. Then, the probability distribution is applied as a constraint in the first Transformer layer to encourage its attention heads to follow the relational pattern structures. As a result, fine-grained pattern information is enhanced in the pre-trained Transformer without losing global dependencies. Extensive experimental results on two popular benchmark datasets demonstrate that our model performs better than the state-of-the-art baselines. (c) 2021 Elsevier Inc. All rights reserved.
引用
收藏
页码:269 / 279
页数:11
相关论文
共 50 条
  • [41] Are Noisy Sentences Useless for Distant Supervised Relation Extraction?
    Shang, Yu-Ming
    Huang, He-Yan
    Mao, Xian-Ling
    Sun, Xin
    Wei, Wei
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 8799 - 8806
  • [42] Weakly supervised action segmentation with effective use of attention and self-attention
    Bin Ng, Yan
    Fernando, Basura
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2021, 213
  • [43] Relevance, valence, and the self-attention network
    Mattan, Bradley D.
    Quinn, Kimberly A.
    Rotshtein, Pia
    COGNITIVE NEUROSCIENCE, 2016, 7 (1-4) : 27 - 28
  • [44] A self-attention network for smoke detection
    Jiang, Minghua
    Zhao, Yaxin
    Yu, Feng
    Zhou, Changlong
    Peng, Tao
    FIRE SAFETY JOURNAL, 2022, 129
  • [45] Aspect term extraction for opinion mining using a Hierarchical Self-Attention Network
    Kumar, Avinash
    Veerubhotla, Aditya Srikanth
    Narapareddy, Vishnu Teja
    Aruru, Vamshi
    Neti, Lalita Bhanu Murthy
    Malapati, Aruna
    NEUROCOMPUTING, 2021, 465 : 195 - 204
  • [46] A time-aware self-attention based neural network model for sequential recommendation
    Zhang, Yihu
    Yang, Bo
    Liu, Haodong
    Li, Dongsheng
    APPLIED SOFT COMPUTING, 2023, 133
  • [47] SAFE: Unsupervised image feature extraction using self-attention based feature extraction network
    Choi, Yeoung Je
    Lee, Gyeong Taek
    Kim, Chang Ouk
    EXPERT SYSTEMS, 2024, 41 (08)
  • [48] Semantic Enhanced Distantly Supervised Relation Extraction via Graph Attention Network
    Ouyang, Xiaoye
    Chen, Shudong
    Wang, Rong
    INFORMATION, 2020, 11 (11) : 1 - 12
  • [49] Relation constraint self-attention for image captioning
    Ji, Junzhong
    Wang, Mingzhan
    Zhang, Xiaodan
    Lei, Minglong
    Qu, Liangqiong
    NEUROCOMPUTING, 2022, 501 : 778 - 789
  • [50] Speaker-Aware Speech Enhancement with Self-Attention
    Lin, Ju
    Van Wijngaarden, Adriaan J.
    Smith, Melissa C.
    Wang, Kuang-Ching
    29TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2021), 2021, : 486 - 490