A Supervised Multi-Head Self-Attention Network for Nested Named Entity Recognition

被引:0
|
作者
Xu, Yongxiu [1 ,2 ]
Huang, Heyan [3 ]
Feng, Chong [3 ]
Hu, Yue [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing, Peoples R China
[3] Beijing Inst Technol, Sch Comp Sci & Technol, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, researchers have shown an increased interest in recognizing the overlapping entities that have nested structures. However, most existing models ignore the semantic correlation between words under different entity types. Considering words in sentence play different roles under different entity types, we argue that the correlation intensities of pairwise words in sentence for each entity type should be considered. In this paper, we treat named entity recognition as a multi-class classification of word pairs and design a simple neural model to handle this issue. Our model applies a supervised multi-head self-attention mechanism, where each head corresponds to one entity type, to construct the word-level correlations for each type. Our model can flexibly predict the span type by the correlation intensities of its head and tail under the corresponding type. In addition, we fuse entity boundary detection and entity classification by a multi-task learning framework, which can capture the dependencies between these two tasks. To verify the performance of our model, we conduct extensive experiments on both nested and flat datasets. The experimental results show that our model can outperform the previous state-of-the-art methods on multiple tasks without any extra NLP tools or human annotations.
引用
收藏
页码:14185 / 14193
页数:9
相关论文
共 50 条
  • [21] Text summarization based on multi-head self-attention mechanism and pointer network
    Qiu, Dong
    Yang, Bing
    COMPLEX & INTELLIGENT SYSTEMS, 2022, 8 (01) : 555 - 567
  • [22] Text summarization based on multi-head self-attention mechanism and pointer network
    Dong Qiu
    Bing Yang
    Complex & Intelligent Systems, 2022, 8 : 555 - 567
  • [23] Self Multi-Head Attention for Speaker Recognition
    India, Miquel
    Safari, Pooyan
    Hernando, Javier
    INTERSPEECH 2019, 2019, : 4305 - 4309
  • [24] A Controlled Attention for Nested Named Entity Recognition
    Chen, Yanping
    Huang, Rong
    Pan, Lijun
    Huang, Ruizhang
    Zheng, Qinghua
    Chen, Ping
    COGNITIVE COMPUTATION, 2023, 15 (01) : 132 - 145
  • [25] A Controlled Attention for Nested Named Entity Recognition
    Yanping Chen
    Rong Huang
    Lijun Pan
    Ruizhang Huang
    Qinghua Zheng
    Ping Chen
    Cognitive Computation, 2023, 15 : 132 - 145
  • [26] Named Entity Recognition in Electronic Medical Records Incorporating Pre-trained and Multi-Head Attention
    Yang, Haotian
    Wang, Li
    Yang, Yanpeng
    IAENG International Journal of Computer Science, 2024, 51 (04) : 401 - 408
  • [27] Adversarial Transfer Learning for Named Entity Recognition Based on Multi-Head Attention Mechanism and Feature Fusion
    Zhao, Dandan
    Zhang, Pan
    Meng, Jiana
    Wu, Yue
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT I, 2022, 13551 : 272 - 284
  • [28] SELF-ATTENTION BASED DARKNET NAMED ENTITY RECOGNITION WITH BERT METHODS
    Chen, Yuxuan
    Guo, Yubin
    Jiang, Hong
    Ding, Jianwei
    Chen, Zhouguo
    INTERNATIONAL JOURNAL OF INNOVATIVE COMPUTING INFORMATION AND CONTROL, 2021, 17 (06): : 1973 - 1988
  • [29] Combined self-attention mechanism for named entity recognition in social media
    Li M.
    Kong F.
    Qinghua Daxue Xuebao/Journal of Tsinghua University, 2019, 59 (06): : 461 - 467
  • [30] Combined Self-Attention Mechanism for Chinese Named Entity Recognition in Military
    Liao, Fei
    Ma, Liangli
    Pei, Jingjing
    Tan, Linshan
    FUTURE INTERNET, 2019, 11 (08):