Leveraging Local and Global Patterns for Self-Attention Networks

被引:0
|
作者
Xu, Mingzhou [1 ]
Wong, Derek F. [1 ]
Yang, Baosong [1 ]
Zhang, Yue [2 ]
Chao, Lidia S. [1 ]
机构
[1] Univ Macau, Dept Comp & Informat Sci, NLP2CT Lab, Taipa, Macao, Peoples R China
[2] Westlake Univ, Sch Engn, Hangzhou, Zhejiang, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Self-attention networks have received increasing research attention. By default, the hidden states of each word are hierarchically calculated by attending to all words in the sentence, which assembles global information. However, several studies pointed out that taking all signals into account may lead to overlooking neighboring information (e.g. phrase pattern). To address this argument, we propose a hybrid attention mechanism to dynamically leverage both of the local and global information. Specifically, our approach uses a gating scalar for integrating both sources of the information, which is also convenient for quantifying their contributions. Experiments on various neural machine translation tasks demonstrate the effectiveness of the proposed method. The extensive analyses verify that the two types of contexts are complementary to each other, and our method gives highly effective improvements in their integration.
引用
收藏
页码:3069 / 3075
页数:7
相关论文
共 50 条
  • [1] On the Global Self-attention Mechanism for Graph Convolutional Networks
    Wang, Chen
    Deng, Chengyuan
    [J]. 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 8531 - 8538
  • [2] Leveraging contextual embeddings and self-attention neural networks with bi-attention for sentiment analysis
    Magdalena Biesialska
    Katarzyna Biesialska
    Henryk Rybinski
    [J]. Journal of Intelligent Information Systems, 2021, 57 : 601 - 626
  • [3] Leveraging contextual embeddings and self-attention neural networks with bi-attention for sentiment analysis
    Biesialska, Magdalena
    Biesialska, Katarzyna
    Rybinski, Henryk
    [J]. JOURNAL OF INTELLIGENT INFORMATION SYSTEMS, 2021, 57 (03) : 601 - 626
  • [4] Convolutional Self-Attention Networks
    Yang, Baosong
    Wang, Longyue
    Wong, Derek F.
    Chao, Lidia S.
    Tu, Zhaopeng
    [J]. 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 4040 - 4045
  • [5] Global-Local Self-Attention Based Transformer for Speaker Verification
    Xie, Fei
    Zhang, Dalong
    Liu, Chengming
    [J]. APPLIED SCIENCES-BASEL, 2022, 12 (19):
  • [6] Leveraging Self-Attention Mechanism for Attitude Estimation in Smartphones
    Brotchie, James
    Shao, Wei
    Li, Wenchao
    Kealy, Allison
    [J]. SENSORS, 2022, 22 (22)
  • [7] Global Convolutional Neural Networks With Self-Attention for Fisheye Image Rectification
    Kim, Byunghyun
    Lee, Dohyun
    Min, Kyeongyuk
    Chong, Jongwha
    Joe, Inwhee
    [J]. IEEE ACCESS, 2022, 10 : 129580 - 129587
  • [8] Global Convolutional Neural Networks With Self-Attention for Fisheye Image Rectification
    Kim, Byunghyun
    Lee, Dohyun
    Min, Kyeongyuk
    Chong, Jongwha
    Joe, Inwhee
    [J]. IEEE Access, 2022, 10 : 129580 - 129587
  • [9] Self-Attention Generative Adversarial Networks
    Zhang, Han
    Goodfellow, Ian
    Metaxas, Dimitris
    Odena, Augustus
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [10] Self-Attention Networks for Code Search
    Fang, Sen
    Tan, You-Shuai
    Zhang, Tao
    Liu, Yepang
    [J]. INFORMATION AND SOFTWARE TECHNOLOGY, 2021, 134