Fine-grained attention mechanism for neural machine translation

被引:122
|
作者
Choi, Heeyoul [1 ]
Cho, Kyunghyun [2 ]
Bengio, Yoshua [3 ]
机构
[1] Handong Global Univ, Pohang, South Korea
[2] NYU, Comp Sci & Data Sci, New York, NY USA
[3] Univ Montreal, Montreal, PQ, Canada
基金
新加坡国家研究基金会; 加拿大自然科学与工程研究理事会;
关键词
Neural machine translation; Attention mechanism; Fine-grained attention;
D O I
10.1016/j.neucom.2018.01.007
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural machine translation (NMT) has been a new paradigm in machine translation, and the attention mechanism has become the dominant approach with the state-of-the-art records in many language pairs. While there are variants of the attention mechanism, all of them use only temporal attention where one scalar value is assigned to one context vector corresponding to a source word. In this paper, we propose a fine-grained (or 2D) attention mechanism where each dimension of a context vector will receive a separate attention score. In experiments with the task of En-De and En-Fi translation, the fine-grained attention method improves the translation quality in terms of BLEU score. In addition, our alignment analysis reveals how the fine-grained attention mechanism exploits the internal structure of context vectors. (C) 2018 Elsevier B.V. All rights reserved.
引用
收藏
页码:171 / 176
页数:6
相关论文
共 50 条
  • [1] Fine-Grained Machine Teaching with Attention Modeling
    Liu, Jiacheng
    Hou, Xiaofeng
    Tang, Feilong
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 2585 - 2592
  • [2] The Lazy Encoder: A Fine-Grained Analysis of the Role of Morphology in Neural Machine Translation
    Bisazza, Arianna
    Tump, Clara
    [J]. 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 2871 - 2876
  • [3] Improving Neural Fine-Grained Entity Typing with Knowledge Attention
    Xin, Ji
    Lin, Yankai
    Liu, Zhiyuan
    Sun, Maosong
    [J]. THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 5997 - 6004
  • [4] A novel fine-grained rumor detection algorithm with attention mechanism
    Zhang, Ke
    Cao, Jianjun
    Pi, Dechang
    [J]. NEUROCOMPUTING, 2024, 583
  • [5] Summary of Fine-Grained Image Recognition Based on Attention Mechanism
    Yao, Ma
    Min, Zhi
    [J]. THIRTEENTH INTERNATIONAL CONFERENCE ON GRAPHICS AND IMAGE PROCESSING (ICGIP 2021), 2022, 12083
  • [6] Attend and Rectify: A Gated Attention Mechanism for Fine-Grained Recovery
    Rodriguez, Pau
    Gonfaus, Josep M.
    Cucurull, Guillem
    Xavier Roca, F.
    Gonzalez, Jordi
    [J]. COMPUTER VISION - ECCV 2018, PT VIII, 2018, 11212 : 357 - 372
  • [7] Pay Attention to the Activations: A Modular Attention Mechanism for Fine-Grained Image Recognition
    Rodriguez, Pau
    Velazquez, Diego
    Cucurull, Guillem
    Gonfaus, Josep M.
    Roca, E. Xavier
    Gonzalez, Jordi
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2020, 22 (02) : 502 - 514
  • [8] FGraDA: A Dataset and Benchmark for Fine-Grained Domain Adaptation in Machine Translation
    Zhu, Wenhao
    Huang, Shujian
    Pu, Tong
    Huang, Pingxuan
    Zhang, Xu
    Yu, Jian
    Chen, Wei
    Wang, Yanfeng
    Chen, Jiajun
    [J]. LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 6719 - 6727
  • [9] Fine-grained image retrieval by combining attention mechanism and context information
    Xiaoqing Li
    Jinwen Ma
    [J]. Neural Computing and Applications, 2023, 35 : 1881 - 1897
  • [10] Fine-grained image retrieval by combining attention mechanism and context information
    Li, Xiaoqing
    Ma, Jinwen
    [J]. NEURAL COMPUTING & APPLICATIONS, 2023, 35 (02): : 1881 - 1897