Self-Training for Unsupervised Neural Machine Translation in Unbalanced Training Data Scenarios

被引:0
|
作者
Sun, Haipeng [1 ,2 ]
Wang, Rui [3 ]
Chen, Kehai [4 ]
Utiyama, Masao [4 ]
Sumita, Eiichiro [4 ]
Zhao, Tiejun [1 ]
机构
[1] Harbin Inst Technol, Harbin, Peoples R China
[2] JD AI Res, Beijing, Peoples R China
[3] Shanghai Jiao Tong Univ, Shanghai, Peoples R China
[4] Natl Inst Informat & Commun Technol NICT, Kyoto, Japan
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Unsupervised neural machine translation (UNMT) that relies solely on massive monolingual corpora has achieved remarkable results in several translation tasks. However, in real-world scenarios, massive monolingual corpora do not exist for some extremely low-resource languages such as Estonian, and UNMT systems usually perform poorly when there is not adequate training corpus for one language. In this paper, we first define and analyze the unbalanced training data scenario for UNMT. Based on this scenario, we propose UNMT self-training mechanisms to train a robust UNMT system and improve its performance in this case. Experimental results on several language pairs show that the proposed methods substantially outperform conventional UNMT systems.
引用
收藏
页码:3975 / 3981
页数:7
相关论文
共 50 条
  • [1] Self-Training Sampling with Monolingual Data Uncertainty for Neural Machine Translation
    Jiao, Wenxiang
    Wang, Xing
    Tu, Zhaopeng
    Shi, Shuming
    Lyu, Michael R.
    King, Irwin
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 2840 - 2850
  • [2] NICT Self-Training Approach to Neural Machine Translation at NMT-2018
    Imamura, Kenji
    Sumita, Eiichiro
    NEURAL MACHINE TRANSLATION AND GENERATION, 2018, : 110 - 115
  • [3] Research on self-training neural machine translation based on monolingual priority sampling
    Zhang X.
    Pang L.
    Du X.
    Lu T.
    Xia Y.
    Tongxin Xuebao/Journal on Communications, 2024, 45 (04): : 65 - 72
  • [4] Unsupervised Controllable Generation with Self-Training
    Chrysos, Grigorios G.
    Kossaifi, Jean
    Yu, Zhiding
    Anandkumar, Anima
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [5] Bridging the Data Gap between Training and Inference for Unsupervised Neural Machine Translation
    He, Zhiwei
    Wang, Xing
    Wang, Rui
    Shi, Shuming
    Tu, Zhaopeng
    PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 6611 - 6623
  • [6] Self-Training for Unsupervised Parsing with PRPN
    Mohananey, Anhad
    Kann, Katharina
    Bowman, Samuel R.
    16TH INTERNATIONAL CONFERENCE ON PARSING TECHNOLOGIES AND IWPT 2020 SHARED TASK ON PARSING INTO ENHANCED UNIVERSAL DEPENDENCIES, 2020, : 105 - 110
  • [7] Iterative Training of Unsupervised Neural and Statistical Machine Translation Systems
    Marie, Benjamin
    Fujita, Atsushi
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2020, 19 (05)
  • [8] Unsupervised Arabic Dialect Adaptation with Self-Training
    Novotney, Scott
    Schwartz, Rich
    Khudanpur, Sanjeev
    12TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2011 (INTERSPEECH 2011), VOLS 1-5, 2011, : 548 - +
  • [9] Attentive continuous generative self-training for unsupervised domain adaptive medical image translation
    Liu, Xiaofeng
    Prince, Jerry L.
    Xing, Fangxu
    Zhuo, Jiachen
    Reese, Timothy
    Stone, Maureen
    El Fakhri, Georges
    Woo, Jonghye
    MEDICAL IMAGE ANALYSIS, 2023, 88
  • [10] Unsupervised domain adaptation with self-training for weed segmentation
    Huang, Yingchao
    Hussein, Amina E.
    Wang, Xin
    Bais, Abdul
    Yao, Shanshan
    Wilder, Tanis
    Intelligent Systems with Applications, 2025, 25