Multi-task Learning for Multilingual Neural Machine Translation

被引:0
|
作者
Wang, Yiren [1 ]
Zhai, ChengXiang [1 ]
Awadalla, Hany Hassan [2 ]
机构
[1] Univ Illinois, Champaign, IL USA
[2] Microsoft, Beijing, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While monolingual data has been shown to be useful in improving bilingual neural machine translation (NMT), effectively and efficiently leveraging monolingual data for Multilingual NMT (MNMT) systems is a less explored area. In this work, we propose a multi-task learning (MTL) framework that jointly trains the model with the translation task on bitext data and two denoising tasks on the monolingual data. We conduct extensive empirical studies on MNMT systems with 10 language pairs from WMT datasets. We show that the proposed approach can effectively improve the translation quality for both high-resource and low-resource languages with large margin, achieving significantly better results than the individual bilingual models. We also demonstrate the efficacy of the proposed approach in the zero-shot setup for language pairs without bitext training data. Furthermore, we show the effectiveness of MTL over pre-training approaches for both NMT and cross-lingual transfer learning NLU tasks; the proposed approach outperforms massive scale models trained on single task.
引用
收藏
页码:1022 / 1034
页数:13
相关论文
共 50 条
  • [1] Improving Robustness of Neural Machine Translation with Multi-task Learning
    Zhou, Shuyan
    Zeng, Xiangkai
    Zhou, Yingqi
    Anastasopoulos, Antonios
    Neubig, Graham
    [J]. FOURTH CONFERENCE ON MACHINE TRANSLATION (WMT 2019), 2019, : 565 - 571
  • [2] Neural Machine Translation Based on Multi-task Learning of Discourse Structure
    Kang X.-M.
    Zong C.-Q.
    [J]. Ruan Jian Xue Bao/Journal of Software, 2022, 33 (10): : 3806 - 3818
  • [3] Training Flexible Depth Model by Multi-Task Learning for Neural Machine Translation
    Wang, Qiang
    Xiao, Tong
    Zhu, Jingbo
    [J]. FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 4307 - 4312
  • [4] Scheduled Multi-task Learning for Neural Chat Translation
    Liang, Yunlong
    Meng, Fandong
    Xu, Jinan
    Chen, Yufeng
    Zhou, Jie
    [J]. PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 4375 - 4388
  • [5] A Multi-task Approach to Learning Multilingual Representations
    Singla, Karan
    Can, Dogan
    Narayanan, Shrikanth
    [J]. PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2, 2018, : 214 - 220
  • [6] Improving Machine Translation of Arabic Dialects Through Multi-task Learning
    Moukafih, Youness
    Sbihi, Nada
    Ghogho, Mounir
    Smaili, Kamel
    [J]. AIXIA 2021 - ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, 13196 : 580 - 590
  • [7] Autocorrect in the Process of Translation- Multi-task Learning Improves Dialogue Machine Translation
    Wang, Tao
    Zhao, Chengqi
    Wang, Mingxuan
    Li, Lei
    Xiong, Deyi
    [J]. 2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, NAACL-HLT 2021, 2021, : 105 - 112
  • [8] Rethinking Data Augmentation for Low-Resource Neural Machine Translation: A Multi-Task Learning Approach
    Sanchez-Cartagena, Victor M.
    Espla-Gomis, Miquel
    Antonio Perez-Ortiz, Juan
    Sanchez-Martinez, Felipe
    [J]. 2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 8502 - 8516
  • [9] Adaptive Knowledge Sharing in Multi-Task Learning: Improving Low-Resource Neural Machine Translation
    Zaremoodi, Poorya
    Buntine, Wray
    Haffari, Gholamreza
    [J]. PROCEEDINGS OF THE 56TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2, 2018, : 656 - 661
  • [10] Multi-Task Learning with Shared Encoder for Non-Autoregressive Machine Translation
    Hao, Yongchang
    He, Shilin
    Jiao, Wenxiang
    Tu, Zhaopeng
    Lyu, Michael R.
    Wang, Xing
    [J]. 2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL-HLT 2021), 2021, : 3989 - 3996