Context-aware positional representation for self-attention networks q

被引:4
|
作者
Chen, Kehai [1 ]
Wang, Rui [1 ]
Utiyama, Masao [1 ]
Sumita, Eiichiro [1 ]
机构
[1] Natl Inst Informat & Commun Technol, Kyoto, Japan
关键词
Positional representation; Context information; Self-attention networks; Machine translation;
D O I
10.1016/j.neucom.2021.04.055
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In self-attention networks (SANs), positional embeddings are used to model order dependencies between words in the input sentence and are added with word embeddings to gain an input representation, which enables the SAN-based neural model to perform (multi-head) and to stack (multi-layer) self-attentive functions in parallel to learn the representation of the input sentence. However, this input representation only involves static order dependencies based on discrete position indexes of words, that is, is independent of context information, which may be weak in modeling the input sentence. To address this issue, we proposed a novel positional representation method to model order dependencies based on n-gram context or sentence context in the input sentence, which allows SANs to learn a more effective sentence representation. To validate the effectiveness of the proposed method, it is applied to the neural machine translation model, which adopts a typical SAN-based neural model. Experimental results on two widely used translation tasks, i.e., WMT14 English-to-German and WMT17 Chinese-to-English, showed that the proposed approach can significantly improve the translation performance over the strong Transformer baseline. (c) 2021 Elsevier B.V. All rights reserved.
引用
下载
收藏
页码:46 / 56
页数:11
相关论文
共 50 条
  • [1] Context-Aware Self-Attention Networks
    Yang, Baosong
    Li, Jian
    Wong, Derek F.
    Chao, Lidia S.
    Wang, Xing
    Tu, Zhaopeng
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 387 - 394
  • [2] Context-aware Self-Attention Networks for Natural Language Processing
    Yang, Baosong
    Wang, Longyue
    Wong, Derek F.
    Shi, Shuming
    Tu, Zhaopeng
    NEUROCOMPUTING, 2021, 458 : 157 - 169
  • [3] Dialogue Act Classification with Context-Aware Self-Attention
    Raheja, Vipul
    Tetreault, Joel
    2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 3727 - 3733
  • [4] Context-Aware Group Captioning via Self-Attention and Contrastive Features
    Li, Zhuowan
    Tran, Quan
    Mai, Long
    Lin, Zhe
    Yuille, Alan L.
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 3437 - 3447
  • [5] Local Context-aware Self-attention for Continuous Sign Language Recognition
    Zuo, Ronglai
    Mak, Brian
    INTERSPEECH 2022, 2022, : 4810 - 4814
  • [6] Spatial Context-Aware Self-Attention Model For Multi-Organ Segmentation
    Tang, Hao
    Liu, Xingwei
    Han, Kun
    Xie, Xiaohui
    Chen, Xuming
    Qian, Huang
    Liu, Yong
    Sun, Shanlin
    Bai, Narisu
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2021), 2021, : 938 - 948
  • [7] A Light-Weight Context-Aware Self-Attention Model for Skin Lesion Segmentation
    Ma, Dongliang
    Wu, Hao
    Sun, Jun
    Yu, Chunjing
    Liu, Li
    PRICAI 2019: TRENDS IN ARTIFICIAL INTELLIGENCE, PT III, 2019, 11672 : 501 - 505
  • [8] Spatial-Temporal Context-Aware Location Prediction Based on Bidirectional Self-Attention Network
    Lin, Kuijie
    Chen, Junxin
    Lian, Xiaoqin
    Mai, Weimin
    Guo, Zhiheng
    Chen, Xiang
    Hsu, Terng-Yin
    2022 14TH INTERNATIONAL CONFERENCE ON WIRELESS COMMUNICATIONS AND SIGNAL PROCESSING, WCSP, 2022, : 701 - 706
  • [9] Context-Aware Semantic Matching with Self Attention Mechanism
    Chen, Yanmin
    Wang, Hao
    Sun, Ruijun
    Chen, Enhong
    2022 5th International Conference on Pattern Recognition and Artificial Intelligence, PRAI 2022, 2022, : 1007 - 1011
  • [10] Sequential Recommendation with Context-Aware Collaborative Graph Attention Networks
    Zhang, Mengfei
    Guo, Cheng
    Jin, Jiaqi
    Pan, Mao
    Fang, Jinyun
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,