Context-Aware Self-Attention Networks

被引:0
|
作者
Yang, Baosong [1 ]
Li, Jian [2 ]
Wong, Derek F. [1 ]
Chao, Lidia S. [1 ]
Wang, Xing [3 ]
Tu, Zhaopeng [3 ]
机构
[1] Univ Macau, Dept Comp & Informat Sci, NLP 2 CT Lab, Macau, Peoples R China
[2] Chinese Univ Hong Kong, Hong Kong, Peoples R China
[3] Tencent AI Lab, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Self-attention model has shown its flexibility in parallel computation and the effectiveness on modeling both long- and short-term dependencies. However, it calculates the dependencies between representations without considering the contextual information, which has proven useful for modeling dependencies among neural representations in various natural language tasks. In this work, we focus on improving self-attention networks through capturing the richness of context. To maintain the simplicity and flexibility of the self-attention networks, we propose to contextualize the transformations of the query and key layers, which are used to calculate the relevance between elements. Specifically, we leverage the internal representations that embed both global and deep contexts, thus avoid relying on external resources. Experimental results on WMT14 English double right arrow German and WMT17 Chinese double right arrow English translation tasks demonstrate the effectiveness and universality of the proposed methods. Furthermore, we conducted extensive analyses to quantify how the context vectors participate in the self-attention model.
引用
收藏
页码:387 / 394
页数:8
相关论文
共 50 条
  • [1] Context-aware positional representation for self-attention networks q
    Chen, Kehai
    Wang, Rui
    Utiyama, Masao
    Sumita, Eiichiro
    NEUROCOMPUTING, 2021, 451 : 46 - 56
  • [2] Context-aware Self-Attention Networks for Natural Language Processing
    Yang, Baosong
    Wang, Longyue
    Wong, Derek F.
    Shi, Shuming
    Tu, Zhaopeng
    NEUROCOMPUTING, 2021, 458 : 157 - 169
  • [3] Dialogue Act Classification with Context-Aware Self-Attention
    Raheja, Vipul
    Tetreault, Joel
    2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 3727 - 3733
  • [4] Context-Aware Group Captioning via Self-Attention and Contrastive Features
    Li, Zhuowan
    Tran, Quan
    Mai, Long
    Lin, Zhe
    Yuille, Alan L.
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 3437 - 3447
  • [5] Local Context-aware Self-attention for Continuous Sign Language Recognition
    Zuo, Ronglai
    Mak, Brian
    INTERSPEECH 2022, 2022, : 4810 - 4814
  • [6] Spatial Context-Aware Self-Attention Model For Multi-Organ Segmentation
    Tang, Hao
    Liu, Xingwei
    Han, Kun
    Xie, Xiaohui
    Chen, Xuming
    Qian, Huang
    Liu, Yong
    Sun, Shanlin
    Bai, Narisu
    2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2021), 2021, : 938 - 948
  • [7] A Light-Weight Context-Aware Self-Attention Model for Skin Lesion Segmentation
    Ma, Dongliang
    Wu, Hao
    Sun, Jun
    Yu, Chunjing
    Liu, Li
    PRICAI 2019: TRENDS IN ARTIFICIAL INTELLIGENCE, PT III, 2019, 11672 : 501 - 505
  • [8] Spatial-Temporal Context-Aware Location Prediction Based on Bidirectional Self-Attention Network
    Lin, Kuijie
    Chen, Junxin
    Lian, Xiaoqin
    Mai, Weimin
    Guo, Zhiheng
    Chen, Xiang
    Hsu, Terng-Yin
    2022 14TH INTERNATIONAL CONFERENCE ON WIRELESS COMMUNICATIONS AND SIGNAL PROCESSING, WCSP, 2022, : 701 - 706
  • [9] Context-Aware Semantic Matching with Self Attention Mechanism
    Chen, Yanmin
    Wang, Hao
    Sun, Ruijun
    Chen, Enhong
    2022 5th International Conference on Pattern Recognition and Artificial Intelligence, PRAI 2022, 2022, : 1007 - 1011
  • [10] Sequential Recommendation with Context-Aware Collaborative Graph Attention Networks
    Zhang, Mengfei
    Guo, Cheng
    Jin, Jiaqi
    Pan, Mao
    Fang, Jinyun
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,