Fast and Robust Neural Network Joint Models for Statistical Machine Translation

被引:0
|
作者
Devlin, Jacob [1 ]
Zbib, Rabih [1 ]
Huang, Zhongqiang [1 ]
Lamar, Thomas [1 ]
Schwartz, Richard [1 ]
Makhoul, John [1 ]
机构
[1] Raytheon BBN Technol, 10 Moulton St, Cambridge, MA 02138 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent work has shown success in using neural network language models (NNLMs) as features in MT systems. Here, we present a novel formulation for a neural network joint model (NNJM), which augments the NNLM with a source context window. Our model is purely lexicalized and can be integrated into any MT decoder. We also present several variations of the NNJM which provide significant additive improvements. Although the model is quite simple, it yields strong empirical results. On the NIST OpenMT12 Arabic-English condition, the NNJM features produce a gain of +3.0 BLEU on top of a powerful, featurerich baseline which already includes a target-only NNLM. The NNJM features also produce a gain of +6.3 BLEU on top of a simpler baseline equivalent to Chiang's (2007) original Hiero implementation. Additionally, we describe two novel techniques for overcoming the historically high cost of using NNLM-style models in MT decoding. These techniques speed up NNJM computation by a factor of 10,000x, making the model as fast as a standard back-off LM.
引用
收藏
页码:1370 / 1380
页数:11
相关论文
共 50 条
  • [21] Recurrent Neural Network based Rule Sequence Model for Statistical Machine Translation
    Yu, Heng
    Zhu, Xuan
    [J]. PROCEEDINGS OF THE 53RD ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL) AND THE 7TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (IJCNLP), VOL 2, 2015, : 132 - 138
  • [22] Bilingual recursive neural network based data selection for statistical machine translation
    Wong, Derek F.
    Lu, Yi
    Chao, Lidia S.
    [J]. KNOWLEDGE-BASED SYSTEMS, 2016, 108 : 15 - 24
  • [23] Lexical Diversity in Statistical and Neural Machine Translation
    Brglez, Mojca
    Vintar, Spela
    [J]. INFORMATION, 2022, 13 (02)
  • [24] Neural and statistical machine translation: perception and productivity
    Lopez Pereira, Ariana
    [J]. TRADUMATICA-TRADUCCIO I TECNOLOGIES DE LA INFORMACIO I LA COMUNICACIO, 2019, (17): : 1 - 19
  • [25] Analysing terminology translation errors in statistical and neural machine translation
    Haque, Rejwanul
    Hasanuzzaman, Mohammed
    Way, Andy
    [J]. MACHINE TRANSLATION, 2020, 34 (2-3) : 149 - 195
  • [26] On the Sparsity of Neural Machine Translation Models
    Wang, Yong
    Wang, Longyue
    Li, Victor O. K.
    Tu, Zhaopeng
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1060 - 1066
  • [27] Robust Neural Machine Translation with ASR Errors
    Xue, Haiyang
    Feng, Yang
    Gu, Shuhao
    Chen, Wei
    [J]. WORKSHOP ON AUTOMATIC SIMULTANEOUS TRANSLATION CHALLENGES, RECENT ADVANCES, AND FUTURE DIRECTIONS, 2020, : 15 - 23
  • [28] Reference Network for Neural Machine Translation
    Fu, Han
    Liu, Chenghao
    Sun, Jianling
    [J]. 57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 3002 - 3012
  • [29] Improving Statistical Machine Translation by Adapting Translation Models to Translationese
    Lembersky, Gennadi
    Ordan, Noam
    Wintner, Shuly
    [J]. COMPUTATIONAL LINGUISTICS, 2013, 39 (04) : 999 - 1024
  • [30] Hybrid Neural Network Alignment and Lexicon Model in Direct HMM for Statistical Machine Translation
    Wang, Weiyue
    Alkhouli, Tamer
    Zhu, Derui
    Ney, Hermann
    [J]. PROCEEDINGS OF THE 55TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2017), VOL 2, 2017, : 125 - 131