Selective Attention for Context-aware Neural Machine Translation

被引:0
|
作者
Maruf, Sameen [1 ]
Martins, Andre F. T. [2 ]
Haffari, Gholamreza [1 ]
机构
[1] Monash Univ, Fac Informat Technol, Melbourne, Vic, Australia
[2] Unbabel Inst Telecomun, Lisbon, Portugal
基金
欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Despite the progress made in sentence-level NMT, current systems still fall short at achieving fluent, good quality translation for a full document. Recent works in context-aware NMT consider only a few previous sentences as context and may not scale to entire documents. To this end, we propose a novel and scalable top-down approach to hierarchical attention for context-aware NMT which uses sparse attention to selectively focus on relevant sentences in the document context and then attends to key words in those sentences. We also propose single-level attention approaches based on sentence or word-level information in the context. The document-level context representation, produced from these attention modules, is integrated into the encoder or decoder of the Transformer model depending on whether we use monolingual or bilingual context. Our experiments and evaluation on English-German datasets in different document MT settings show that our selective attention approach not only significantly outperforms context-agnostic baselines but also surpasses context-aware baselines in most cases.
引用
收藏
页码:3092 / 3102
页数:11
相关论文
共 50 条
  • [41] Context-aware pyramid attention network for crowd counting
    Lingyu Gu
    Chen Pang
    Yanjun Zheng
    Chen Lyu
    Lei Lyu
    [J]. Applied Intelligence, 2022, 52 : 6164 - 6180
  • [42] Context-aware pyramid attention network for crowd counting
    Gu, Lingyu
    Pang, Chen
    Zheng, Yanjun
    Lyu, Chen
    Lyu, Lei
    [J]. APPLIED INTELLIGENCE, 2022, 52 (06) : 6164 - 6180
  • [43] Graph Neural Network for Context-Aware Recommendation
    Sattar, Asma
    Bacciu, Davide
    [J]. NEURAL PROCESSING LETTERS, 2023, 55 (05) : 5357 - 5376
  • [44] Graph Neural Network for Context-Aware Recommendation
    Asma Sattar
    Davide Bacciu
    [J]. Neural Processing Letters, 2023, 55 : 5357 - 5376
  • [45] CONTEXT-AWARE ATTENTION MECHANISM FOR SPEECH EMOTION RECOGNITION
    Ramet, Gaetan
    Garner, Philip N.
    Baeriswyl, Michael
    Lazaridis, Alexandros
    [J]. 2018 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY (SLT 2018), 2018, : 126 - 131
  • [46] Graph Attention Network for Context-Aware Visual Tracking
    Shao, Yanyan
    Guo, Dongyan
    Cui, Ying
    Wang, Zhenhua
    Zhang, Liyan
    Zhang, Jianhua
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024,
  • [47] Context-Aware Semantic Matching with Self Attention Mechanism
    Chen, Yanmin
    Wang, Hao
    Sun, Ruijun
    Chen, Enhong
    [J]. 2022 5th International Conference on Pattern Recognition and Artificial Intelligence, PRAI 2022, 2022, : 1007 - 1011
  • [48] Neural Machine Translation with Deep Attention
    Zhang, Biao
    Xiong, Deyi
    Su, Jinsong
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2020, 42 (01) : 154 - 163
  • [49] Recurrent Attention for Neural Machine Translation
    Zeng, Jiali
    Wu, Shuangzhi
    Yin, Yongjing
    Jiang, Yufan
    Li, Mu
    [J]. 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 3216 - 3225
  • [50] Graph convolution machine for context-aware recommender system
    Jiancan Wu
    Xiangnan He
    Xiang Wang
    Qifan Wang
    Weijian Chen
    Jianxun Lian
    Xing Xie
    [J]. Frontiers of Computer Science, 2022, 16