Multi-Source Neural Machine Translation With Missing Data

被引:11
|
作者
Nishimura, Yuta [1 ]
Sudoh, Katsuhito [1 ]
Neubig, Graham [2 ]
Nakamura, Satoshi [1 ]
机构
[1] Nara Inst Sci & Technol, Ikoma 6300192, Japan
[2] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
关键词
Neural machine translation (NMT); multi-linguality; data augmentation;
D O I
10.1109/TASLP.2019.2959224
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Machine translation is rife with ambiguities in word ordering and word choice, and even with the advent of machine-learning methods that learn to resolve this ambiguity based on statistics from large corpora, mistakes are frequent. Multi-source translation is an approach that attempts to resolve these ambiguities by exploiting multiple inputs (e.g. sentences in three different languages) to increase translation accuracy. These methods are trained on multilingual corpora, which include the multiple source languages and the target language, and then at test time uses information from both source languages while generating the target. While there are many of these multilingual corpora, such as multilingual translations of TED talks or European parliament proceedings, in practice, many multilingual corpora are not complete due to the difficulty to provide translations in all of the relevant languages. Existing studies on multi-source translation did not explicitly handle such situations, and thus are only applicable to complete corpora that have all of the languages of interest, severely limiting their practical applicability. In this article, we examine approaches for multi-source neural machine translation (NMT) that can learn from and translate such incomplete corpora. Specifically, we propose methods to deal with incomplete corpora at both training time and test time. For training time, we examine two methods: (1) a simple method that simply replaces missing source translations with a special NULL symbol, and (2) a data augmentation approach that fills in incomplete parts with source translations created from multi-source NMT. For test-time, we examine methods that use multi-source translation even when only a single source is provided by first translating into an additional auxiliary language using standard NMT, then using multi-source translation on the original source and this generated auxiliary language sentence. Extensive experiments demonstrate that the proposed training-time and test-time methods both significantly improve translation performance.
引用
收藏
页码:569 / 580
页数:12
相关论文
共 50 条
  • [1] Multi-Source Neural Machine Translation with Missing Data
    Nishimura, Yuta
    Sudoh, Katsuhito
    Neubig, Graham
    Nakamura, Satoshi
    [J]. NEURAL MACHINE TRANSLATION AND GENERATION, 2018, : 92 - 99
  • [2] Multi-Source Syntactic Neural Machine Translation
    Currey, Anna
    Heafield, Kenneth
    [J]. 2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 2961 - 2966
  • [3] Ensemble learning for multi-source neural machine translation
    [J]. 1600, Association for Computational Linguistics, ACL Anthology
  • [4] Multi-Source Neural Model for Machine Translation of Agglutinative Language
    Pan, Yirong
    Li, Xiao
    Yang, Yating
    Dong, Rui
    [J]. FUTURE INTERNET, 2020, 12 (06):
  • [5] Exploring Multi-Stage Information Interactions for Multi-Source Neural Machine Translation
    Lu, Ziyao
    Li, Xiang
    Liu, Yang
    Zhou, Chulun
    Cui, Jianwei
    Wang, Bin
    Zhang, Min
    Su, Jinsong
    [J]. IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2022, 30 : 562 - 570
  • [6] Multi-Source Transformer for Kazakh-Russian-English Neural Machine Translation
    Littell, Patrick
    Lo, Chi-kiu
    Larkin, Samuel
    Stewart, Darlene
    [J]. FOURTH CONFERENCE ON MACHINE TRANSLATION (WMT 2019), 2019, : 267 - 274
  • [7] Multi-source Data Hiding in Neural Networks
    Yang, Ziyun
    Wang, Zichi
    Zhang, Xinpeng
    Tang, Zhenjun
    [J]. 2022 IEEE 24TH INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP), 2022,
  • [8] Improving a Multi-Source Neural Machine Translation Model with Corpus Extension for Low-Resource Languages
    Choi, Gyu-Hyeon
    Shin, Jong-Hun
    Kim, Young-Kil
    [J]. PROCEEDINGS OF THE ELEVENTH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2018), 2018, : 900 - 904
  • [9] Cheat Codes to Quantify Missing Source Information in Neural Machine Translation
    Pal, Proyag
    Heafield, Kenneth
    [J]. NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 2472 - 2477
  • [10] Neural TV program recommendation with multi-source heterogeneous data
    Yin, Fulian
    Xing, Tongtong
    Wu, Zhaoliang
    Feng, Xiaoli
    Ji, Meiqi
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 119