Capsule Network Improved Multi-Head Attention for Word Sense Disambiguation

被引:0
|
作者
Cheng, Jinfeng [1 ]
Tong, Weiqin [1 ,2 ]
Yan, Weian [1 ]
机构
[1] Shanghai Univ, Sch Comp Engn & Sci, Shanghai 200444, Peoples R China
[2] Shanghai Univ, Shanghai Inst Adv Commun & Data Sci, Shanghai 200444, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2021年 / 11卷 / 06期
关键词
word sense disambiguation; multi-head attention; capsule network; capsule routing;
D O I
10.3390/app11062488
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Word sense disambiguation (WSD) is one of the core problems in natural language processing (NLP), which is to map an ambiguous word to its correct meaning in a specific context. There has been a lively interest in incorporating sense definition (gloss) into neural networks in recent studies, which makes great contribution to improving the performance of WSD. However, disambiguating polysemes of rare senses is still hard. In this paper, while taking gloss into consideration, we further improve the performance of the WSD system from the perspective of semantic representation. We encode the context and sense glosses of the target polysemy independently using encoders with the same structure. To obtain a better presentation in each encoder, we leverage the capsule network to capture different important information contained in multi-head attention. We finally choose the gloss representation closest to the context representation of the target word as its correct sense. We do experiments on English all-words WSD task. Experimental results show that our method achieves good performance, especially having an inspiring effect on disambiguating words of rare senses.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Multi-Head Self-Attention Gated-Dilated Convolutional Neural Network for Word Sense Disambiguation
    Zhang, Chun-Xiang
    Zhang, Yu-Long
    Gao, Xue-Yao
    IEEE ACCESS, 2023, 11 : 14202 - 14210
  • [2] Improving Multi-head Attention with Capsule Networks
    Gu, Shuhao
    Feng, Yang
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING (NLPCC 2019), PT I, 2019, 11838 : 314 - 326
  • [3] Attention Neural Network for Biomedical Word Sense Disambiguation
    Zhang, Chun-Xiang
    Pang, Shu-Yang
    Gao, Xue-Yao
    Lu, Jia-Qi
    Yu, Bo
    DISCRETE DYNAMICS IN NATURE AND SOCIETY, 2022, 2022
  • [4] Acoustic Word Embedding Based on Multi-Head Attention Quadruplet Network
    Zhu, Shirong
    Zhang, Ying
    He, Kai
    Zhao, Lasheng
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 184 - 188
  • [5] Word embedding factor based multi-head attention
    Li, Zhengren
    Zhao, Yumeng
    Zhang, Xiaohang
    Han, Huawei
    Huang, Cui
    ARTIFICIAL INTELLIGENCE REVIEW, 2025, 58 (04)
  • [6] An Improved Word Sense Disambiguation Method
    Yu, Linlin
    Song, Lifang
    Sun, Jianyan
    Li, Lin
    2016 6TH INTERNATIONAL CONFERENCE ON INFORMATION TECHNOLOGY FOR MANUFACTURING SYSTEMS (ITMS 2016), 2016, : 153 - 155
  • [7] An Improved Approach to Word Sense Disambiguation
    Sachdeva, Pradeep
    Verma, Surbhi
    Singh, Sandeep Kumar
    2014 IEEE INTERNATIONAL SYMPOSIUM ON SIGNAL PROCESSING AND INFORMATION TECHNOLOGY (ISSPIT), 2014, : 235 - 240
  • [8] An improved algorithm on word sense disambiguation
    Serban, G
    Tatar, D
    INTELLIGENT INFORMATION PROCESSING AND WEB MINING, 2003, : 199 - 208
  • [9] An interactive multi-head self-attention capsule network model for aspect sentiment classification
    She, Lina
    Gong, Hongfang
    Zhang, Siyu
    JOURNAL OF SUPERCOMPUTING, 2024, 80 (07): : 9327 - 9352
  • [10] An interactive multi-head self-attention capsule network model for aspect sentiment classification
    Lina She
    Hongfang Gong
    Siyu Zhang
    The Journal of Supercomputing, 2024, 80 : 9327 - 9352