Word-Based Domain Feature-Sensitive Multi-domain Neural Machine Translation

被引:0
|
作者
Huang Z. [1 ]
Man Z. [1 ]
Zhang Y. [1 ]
Xu J. [1 ]
Chen Y. [1 ]
机构
[1] School of Computer and Information Technology, Beijing Jiaotong University, Beijing
基金
美国国家卫生研究院;
关键词
context features; domain discrimination; domain feature-sensitive; multi-domain NMT;
D O I
10.13209/j.0479-8023.2022.063
中图分类号
TN911 [通信理论];
学科分类号
081002 ;
摘要
The accuracy of the existing word-based domain feature learning methods on domain discrimination is still low and the further research for domain feature learning is required. In order to improve domain discrimination and provide accurate translation, this paper proposes a word-based domain feature-sensitive learning mechanism, including 1) the context feature encoding at encoder side, to widen the study range of word-based domain features, introducing convolutional neural networks (CNN) in encoder for extracting features from word strings with different lengths in parallel as word context features; and 2) enhanced domain feature learning. A domain discriminator module based on multi-layer perceptions (MLP) is designed to enhance the learning ability of obtaining more accurate domain proportion from word context features and improve the accuracy of word domain discrimination. Experiments on English-Chinese task of UM-Corpus and English-French task of OPUS show that the average BLEU scores of the proposed method exceed the strong baseline by 0.82 and 1.06 respectively. The accuracy of domain discrimination is improved by 10.07% and 18.06% compared with the baseline. More studies illustrate that the improvements of average BLEU scores and accuracy of domain discrimation are contributed by the proposed word-based domain feature-sensitive learning mechanism. © 2023 Peking University. All rights reserved.
引用
收藏
页码:1 / 10
页数:9
相关论文
共 21 条
  • [1] Sutskever I, Vinyals O, Le Q V., Sequence to sequence learning with neural networks, NIPS, pp. 3104-3112, (2014)
  • [2] Bahdanau D, Cho K, Bengio Y., Neural machine translation by jointly learning to align and translate, ICLR, pp. 1-15, (2015)
  • [3] Luong M T, Pham H, Manning C D., Effective approaches to attention-based neural machine translation, EMNLP, pp. 1412-1421, (2015)
  • [4] Vaswani A, Shazeer N, Parmar N, Et al., Attention is all you need // NIPS, pp. 5998-6008, (2017)
  • [5] Sajjad H, Durrani N, Dalvi F, Et al., Neural machine translation training in a multi-domain scenario, Proceedings of the 14th International Conference on Spoken Language Translation, pp. 66-73, (2017)
  • [6] Haddow B, Koehn P., Analysing the effect of out-of-domain data on SMT systems, Proceedings of the Seventh Workshop on Statistical Machine Translation, pp. 422-432, (2012)
  • [7] Britz D, Le Q, Pryzant R., Effective domain mixing for neural machine translation, Proceedings of the Second Conference on Machine Translation, pp. 118-126, (2017)
  • [8] Jiali Zeng, Jinsong Su, Huating Wen, Et al., Multi-domain neural machine translation with word-level domain context discrimination, EMNLP, pp. 447-457, (2018)
  • [9] Haoming Jiang, Chen Liang, Wang Chong, Et al., Multi-domain neural machine translation with word-level adaptive layer-wise domain mixing [C/OL], Proceedings of the ACL
  • [10] Kobus C, Crego J, Senellart J., Domain control for neural machine translation, Proceedings of the International Conference Recent Advances in Natural Language Processing, pp. 372-378, (2017)