Self-attention enhanced CNNs with average margin loss for chinese zero pronoun resolution

被引:2
|
作者
Sun, Shi-jun [1 ]
机构
[1] Shanghai Univ, Inst Comp Engn & Sci, 99 Shangda Rd, Shanghai 200444, Peoples R China
关键词
Zero pronoun resolution; Self-attention; Convolutional neural networks; Natural language processing;
D O I
10.1007/s10489-021-02697-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent neural network methods for Chinese zero pronoun resolution explore multiple models for generating representation vectors for zero pronouns and their candidate antecedents. Typically, the representation of zero pronouns are generated by their contextual information since they are simply gaps, which makes it hard to express them. To better interpret zero pronouns and their candidate antecedents, we here introduce a convolutional neural networks with internal self-attention module for encoding them. With the help of the Multi-hop attention mechanism, our model is able to focus on informative parts of the associated contextual texts, which produces an effective way to capture these important information. In addition, we propose a novel average margin loss by averaging the candidate antecedents scores, making the model learning more reasonable. Experimental results on OntoNotes 5.0 dataset show that our model gains the best performance on the task of Chinese zero pronoun resolution.
引用
收藏
页码:5739 / 5750
页数:12
相关论文
共 50 条
  • [31] Contextualized dynamic meta embeddings based on Gated CNNs and self-attention for Arabic machine translation
    Bensalah, Nouhaila
    Ayad, Habib
    Adib, Abdellah
    El Farouk, Abdelhamid Ibn
    INTERNATIONAL JOURNAL OF INTELLIGENT COMPUTING AND CYBERNETICS, 2024, 17 (03) : 605 - 631
  • [32] Chinese Zero Pronoun Resolution: An Unsupervised Approach Combining Ranking and Integer Linear Programming
    Chen, Chen
    Ng, Vincent
    PROCEEDINGS OF THE TWENTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2014, : 1622 - 1628
  • [33] SELF-ATTENTION AND RETRIEVAL ENHANCED NEURAL NETWORKS FOR ESSAY GENERATION
    Wang, Wei
    Zheng, Hai-Tao
    Lin, Zibo
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 8199 - 8203
  • [34] Self-Attention Enhanced Recurrent Neural Networks for Sentence Classification
    Kumar, Ankit
    Rastogi , Reshma
    2018 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI), 2018, : 905 - 911
  • [35] Self-attention Enhanced Patient Journey Understanding in Healthcare System
    Peng, Xueping
    Long, Guodong
    Shen, Tao
    Wang, Sen
    Jiang, Jing
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2020, PT III, 2021, 12459 : 719 - 735
  • [36] Lightweight Vision Transformer with Spatial and Channel Enhanced Self-Attention
    Zheng, Jiahao
    Yang, Longqi
    Li, Yiying
    Yang, Ke
    Wang, Zhiyuan
    Zhou, Jun
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 1484 - 1488
  • [37] Multi-feature self-attention super-resolution network
    Yang, Aiping
    Wei, Zihao
    Wang, Jinbin
    Cao, Jiale
    Ji, Zhong
    Pang, Yanwei
    VISUAL COMPUTER, 2024, 40 (05): : 3473 - 3486
  • [38] Multi-feature self-attention super-resolution network
    Aiping Yang
    Zihao Wei
    Jinbin Wang
    Jiale Cao
    Zhong Ji
    Yanwei Pang
    The Visual Computer, 2024, 40 : 3473 - 3486
  • [39] SRFormer: Permuted Self-Attention for Single Image Super-Resolution
    Zhou, Yupeng
    Li, Zhen
    Guo, Chun-Le
    Bai, Song
    Cheng, Ming-Ming
    Hou, Qibin
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 12734 - 12745
  • [40] Enhanced self-supervised monocular depth estimation with self-attention and joint depth-pose loss for laparoscopic images
    Li, Wenda
    Hayashi, Yuichiro
    Oda, Masahiro
    Kitasaka, Takayuki
    Misawa, Kazunari
    Mori, Kensaku
    INTERNATIONAL JOURNAL OF COMPUTER ASSISTED RADIOLOGY AND SURGERY, 2025,