Low-Resource Neural Machine Translation Using XLNet Pre-training Model

被引:1
|
作者
Wu, Nier [1 ]
Hou, Hongxu [1 ]
Guo, Ziyue [1 ]
Zheng, Wei [1 ]
机构
[1] Inner Mongolia Univ, Coll Comp Sci, Coll Software, Hohhot, Inner Mongolia, Peoples R China
关键词
Low-resource; Machine translation; XLNet; Pre-training;
D O I
10.1007/978-3-030-86383-8_40
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The methods to improve the quality of low-resource neural machine translation (NMT) include: change the token granularity to reduce the number of low-frequency words; generate pseudo-parallel corpus from large-scale monolingual data to optimize model parameters; Use the auxiliary knowledge of pre-trained model to train NMT model. However, reducing token granularity will result in a large number of invalid operations and increase the complexity of local reordering on the target side. Pseudo-parallel corpus contains noise affect model convergence. Pre-training methods also limit translation quality due to the human error and the assumption of conditional independence. Therefore, we proposed a XLNet based pre-training method, that corrects the defects of the pre-training model, and enhance NMT model for context feature extraction. Experiments are carried out on CCMT2019 Mongolian-Chinese (Mo-Zh), Uyghur-Chinese (Ug-Zh) and Tibetan-Chinese (Ti-Zh) tasks, the results show that the generalization ability and BLEU scores of our method are improved compared with the baseline, which fully verifies the effectiveness of the method.
引用
收藏
页码:503 / 514
页数:12
相关论文
共 50 条
  • [41] Augmenting training data with syntactic phrasal-segments in low-resource neural machine translation
    Gupta, Kamal Kumar
    Sen, Sukanta
    Haque, Rejwanul
    Ekbal, Asif
    Bhattacharyya, Pushpak
    Way, Andy
    [J]. MACHINE TRANSLATION, 2021, 35 (04) : 661 - 685
  • [42] Survey of Low-Resource Machine Translation
    Haddow, Barry
    Bawden, Rachel
    Barone, Antonio Valerio Miceli
    Helcl, Jindrich
    Birch, Alexandra
    [J]. COMPUTATIONAL LINGUISTICS, 2022, 48 (03) : 673 - 732
  • [43] Neural machine translation of low-resource languages using SMT phrase pair injection
    Sen, Sukanta
    Hasanuzzaman, Mohammed
    Ekbal, Asif
    Bhattacharyya, Pushpak
    Way, Andy
    [J]. NATURAL LANGUAGE ENGINEERING, 2021, 27 (03) : 271 - 292
  • [44] LenM: Improving Low-Resource Neural Machine Translation Using Target Length Modeling
    Mahsuli, Mohammad Mahdi
    Khadivi, Shahram
    Homayounpour, Mohammad Mehdi
    [J]. NEURAL PROCESSING LETTERS, 2023, 55 (07) : 9435 - 9466
  • [45] Multimodal Neural Machine Translation for Low-resource Language Pairs using Synthetic Data
    Chowdhury, Koel Dutta
    Hasanuzzaman, Mohammed
    Liu, Qun
    [J]. DEEP LEARNING APPROACHES FOR LOW-RESOURCE NATURAL LANGUAGE PROCESSING (DEEPLO), 2018, : 33 - 42
  • [46] LenM: Improving Low-Resource Neural Machine Translation Using Target Length Modeling
    Mohammad Mahdi Mahsuli
    Shahram Khadivi
    Mohammad Mehdi Homayounpour
    [J]. Neural Processing Letters, 2023, 55 : 9435 - 9466
  • [47] Simulated Multiple Reference Training Improves Low-Resource Machine Translation
    Khayrallah, Huda
    Thompson, Brian
    Post, Matt
    Koehn, Philipp
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 82 - 89
  • [48] Rethinking the Exploitation of Monolingual Data for Low-Resource Neural Machine Translation
    Pang, Jianhui
    Yang, Baosong
    Wong, Derek Fai
    Wan, Yu
    Liu, Dayiheng
    Chao, Lidia Sam
    Xie, Jun
    [J]. COMPUTATIONAL LINGUISTICS, 2023, 50 (01) : 25 - 47
  • [49] A Diverse Data Augmentation Strategy for Low-Resource Neural Machine Translation
    Li, Yu
    Li, Xiao
    Yang, Yating
    Dong, Rui
    [J]. INFORMATION, 2020, 11 (05)
  • [50] Semantic Perception-Oriented Low-Resource Neural Machine Translation
    Wu, Nier
    Hou, Hongxu
    Li, Haoran
    Chang, Xin
    Jia, Xiaoning
    [J]. MACHINE TRANSLATION, CCMT 2021, 2021, 1464 : 51 - 62