A Supervised Multi-Head Self-Attention Network for Nested Named Entity Recognition

被引:0
|
作者
Xu, Yongxiu [1 ,2 ]
Huang, Heyan [3 ]
Feng, Chong [3 ]
Hu, Yue [1 ,2 ]
机构
[1] Chinese Acad Sci, Inst Informat Engn, Beijing, Peoples R China
[2] Univ Chinese Acad Sci, Sch Cyber Secur, Beijing, Peoples R China
[3] Beijing Inst Technol, Sch Comp Sci & Technol, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, researchers have shown an increased interest in recognizing the overlapping entities that have nested structures. However, most existing models ignore the semantic correlation between words under different entity types. Considering words in sentence play different roles under different entity types, we argue that the correlation intensities of pairwise words in sentence for each entity type should be considered. In this paper, we treat named entity recognition as a multi-class classification of word pairs and design a simple neural model to handle this issue. Our model applies a supervised multi-head self-attention mechanism, where each head corresponds to one entity type, to construct the word-level correlations for each type. Our model can flexibly predict the span type by the correlation intensities of its head and tail under the corresponding type. In addition, we fuse entity boundary detection and entity classification by a multi-task learning framework, which can capture the dependencies between these two tasks. To verify the performance of our model, we conduct extensive experiments on both nested and flat datasets. The experimental results show that our model can outperform the previous state-of-the-art methods on multiple tasks without any extra NLP tools or human annotations.
引用
收藏
页码:14185 / 14193
页数:9
相关论文
共 50 条
  • [41] An interactive multi-head self-attention capsule network model for aspect sentiment classification
    She, Lina
    Gong, Hongfang
    Zhang, Siyu
    JOURNAL OF SUPERCOMPUTING, 2024, 80 (07): : 9327 - 9352
  • [42] GlobalMind: Global multi-head interactive self-attention network for hyperspectral change detection
    Hu, Meiqi
    Wu, Chen
    Zhang, Liangpei
    ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2024, 211 : 465 - 483
  • [43] An interactive multi-head self-attention capsule network model for aspect sentiment classification
    Lina She
    Hongfang Gong
    Siyu Zhang
    The Journal of Supercomputing, 2024, 80 : 9327 - 9352
  • [44] Drug-Target Interaction Prediction Using Multi-Head Self-Attention and Graph Attention Network
    Cheng, Zhongjian
    Yan, Cheng
    Wu, Fang-Xiang
    Wang, Jianxin
    IEEE-ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS, 2022, 19 (04) : 2208 - 2218
  • [45] Adaptive Multi-Head Self-Attention Based supervised VAE for Industrial Soft Sensing With Missing Data
    Chen, Lei
    Xu, Yuan
    Zhu, Qun-Xiong
    He, Yan-Lin
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2024, 21 (03) : 3564 - 3575
  • [46] A self-attention based neural architecture for Chinese medical named entity recognition
    Wan, Qian
    Liu, Jie
    Wei, Luona
    Ji, Bin
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2020, 17 (04) : 3498 - 3511
  • [47] Named Entity Recognition of BERT-BiLSTM-CRF Combined with Self-attention
    Xu, Lei
    Li, Shuang
    Wang, Yuchen
    Xu, Lizhen
    WEB INFORMATION SYSTEMS AND APPLICATIONS (WISA 2021), 2021, 12999 : 556 - 564
  • [48] Uniting Multi-Scale Local Feature Awareness and the Self-Attention Mechanism for Named Entity Recognition
    Shi, Lin
    Zou, Xianming
    Dai, Chenxu
    Ji, Zhanlin
    MATHEMATICS, 2023, 11 (11)
  • [49] Multidimensional Self-Attention for Aspect Term Extraction and Biomedical Named Entity Recognition
    Song, Xinyu
    Feng, Ao
    Wang, Weikuan
    Gao, Zhengjie
    MATHEMATICAL PROBLEMS IN ENGINEERING, 2020, 2020
  • [50] Adversarial Transfer Learning for Chinese Named Entity Recognition with Self-Attention Mechanism
    Cao, Pengfei
    Chen, Yubo
    Liu, Kang
    Zhao, Jun
    Liu, Shengping
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 182 - 192