Enhancing Machine Translation with Dependency-Aware Self-Attention

被引:0
|
作者
Bugliarello, Emanuele [1 ,2 ]
Okazaki, Naoaki [2 ]
机构
[1] Univ Copenhagen, Copenhagen, Denmark
[2] Tokyo Inst Technol, Tokyo, Japan
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Most neural machine translation models only rely on pairs of parallel sentences, assuming syntactic information is automatically learned by an attention mechanism. In this work, we investigate different approaches to incorporate syntactic knowledge in the Transformer model and also propose a novel, parameter-free, dependency-aware self-attention mechanism that improves its translation quality, especially for long sentences and in low-resource scenarios. We show the efficacy of each approach on WMT English <-> German and English -> Turkish, and WAT English -> Japanese translation tasks.
引用
收藏
页码:1618 / 1627
页数:10
相关论文
共 50 条
  • [1] Domain-Aware Self-Attention for Multi-Domain Neural Machine Translation
    Zhang, Shiqi
    Liu, Yan
    Xiong, Deyi
    Zhang, Pei
    Chen, Boxing
    [J]. INTERSPEECH 2021, 2021, : 2047 - 2051
  • [2] Toward Dependency-Aware Live Virtual Machine Migration
    Nocentino, Anthony
    Ruth, Paul M.
    [J]. THIRD INTERNATIONAL WORKSHOP ON VIRTUALIZATION TECHNOLOGIES IN DISTRIBUTED COMPUTING (VTDC-09), 2009, : 59 - 66
  • [3] Dependency-Aware Attention Model for Emotion Analysis for Online News
    Zhao, Xue
    Zhang, Ying
    Yuan, Xiaojie
    [J]. ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2019, PT I, 2019, 11439 : 172 - 184
  • [4] Multi-Granularity Self-Attention for Neural Machine Translation
    Hao, Jie
    Wang, Xing
    Shi, Shuming
    Zhang, Jinfeng
    Tu, Zhaopeng
    [J]. 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 887 - 897
  • [5] Enhancing low-resource neural machine translation with syntax-graph guided self-attention
    Gong, Longchao
    Li, Yan
    Guo, Junjun
    Yu, Zhengtao
    Gao, Shengxiang
    [J]. KNOWLEDGE-BASED SYSTEMS, 2022, 246
  • [6] Dependency-aware Form Understanding
    Zhang, Shaokun
    Li, Yuanchun
    Yan, Weixiang
    Guo, Yao
    Chen, Xiangqun
    [J]. 2021 IEEE 32ND INTERNATIONAL SYMPOSIUM ON SOFTWARE RELIABILITY ENGINEERING (ISSRE 2021), 2021, : 139 - 149
  • [7] Dependency-Aware Attention Control for Unconstrained Face Recognition with Image Sets
    Liu, Xiaofeng
    Kumar, B. V. K. Vijaya
    Yang, Chao
    Tang, Qingming
    You, Jane
    [J]. COMPUTER VISION - ECCV 2018, PT XI, 2018, 11215 : 573 - 590
  • [8] Why Self-Attention? A Targeted Evaluation of Neural Machine Translation Architectures
    Tang, Gongbo
    Mueller, Mathias
    Rios, Annette
    Sennrich, Rico
    [J]. 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 4263 - 4272
  • [9] Re-Transformer: A Self-Attention Based Model for Machine Translation
    Liu, Huey-Ing
    Chen, Wei-Lin
    [J]. AI IN COMPUTATIONAL LINGUISTICS, 2021, 189 : 3 - 10
  • [10] English Machine Translation Model Based on an Improved Self-Attention Technology
    Pan, Wenxia
    [J]. SCIENTIFIC PROGRAMMING, 2021, 2021