Multilingual Pre-training Model-Assisted Contrastive Learning Neural Machine Translation

被引:0
|
作者
Sun, Shuo [1 ]
Hou, Hong-xu [1 ]
Yang, Zong-heng [1 ]
Wang, Yi-song [1 ]
机构
[1] Inner Mongolia Univ, Coll Comp Sci, Natl & Local Joint Engn Res Ctr Intelligent Infor, Inner Mongolia Key Lab Mongolian Informat Proc Te, Hohhot, Peoples R China
关键词
Low-Resource NMT; Pre-training Model; Contrastive Learning; Dynamic Training;
D O I
10.1109/IJCNN54540.2023.10191766
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Since pre-training and fine-tuning have been a successful paradigm in Natural Language Processing (NLP), this paper adopts the SOTA pre-training model-CeMAT as a strong assistant for low-resource ethnic language translation tasks. Aiming at the exposure bias problem in the fine-tuning process, we use the contrastive learning framework and propose a new contrastive examples generation method, which uses self-generated predictions as contrastive examples to expose the model to errors during inference. Moreover, in order to effectively utilize the limited bilingual data in low-resource tasks, this paper proposes a dynamic training strategy to fine-tune the model, and refines the model step by step by taking word embedding norm and uncertainty as the criteria of evaluate data and model respectively. Experimental results demonstrate that our method significantly improves the quality compared to the baselines, which fully verifies the effectiveness.
引用
收藏
页数:7
相关论文
共 50 条
  • [21] SemFace: Pre-training Encoder and Decoder with a Semantic Interface for Neural Machine Translation
    Ren, Shuo
    Zhou, Long
    Liu, Shujie
    Wei, Furu
    Zhou, Ming
    Ma, Shuai
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 4518 - 4527
  • [22] Pre-Training on Mixed Data for Low-Resource Neural Machine Translation
    Zhang, Wenbo
    Li, Xiao
    Yang, Yating
    Dong, Rui
    INFORMATION, 2021, 12 (03)
  • [23] Soft Language Clustering for Multilingual Model Pre-training
    Zeng, Jiali
    Jiang, Yufan
    Yin, Yongjing
    Jing, Yi
    Meng, Fandong
    Lin, Binghuai
    Cao, Yunbo
    Zhou, Jie
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL 2023, VOL 1, 2023, : 7021 - 7035
  • [24] PoisonedEncoder: Poisoning the Unlabeled Pre-training Data in Contrastive Learning
    Liu, Hongbin
    Jia, Jinyuan
    Gong, Neil Zhenqiang
    PROCEEDINGS OF THE 31ST USENIX SECURITY SYMPOSIUM, 2022, : 3629 - 3645
  • [25] A Contrastive Learning Pre-Training Method for Motif Occupancy Identification
    Lin, Ken
    Quan, Xiongwen
    Yin, Wenya
    Zhang, Han
    INTERNATIONAL JOURNAL OF MOLECULAR SCIENCES, 2022, 23 (09)
  • [26] JASS: Japanese-specific Sequence to Sequence Pre-training for Neural Machine Translation
    Mao, Zhuoyuan
    Cromieres, Fabien
    Dabre, Raj
    Song, Haiyue
    Kurohashi, Sadao
    PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020), 2020, : 3683 - 3691
  • [27] Active Learning with Contrastive Pre-training for Facial Expression Recognition
    Roy, Shuvendu
    Etemad, Ali
    2023 11TH INTERNATIONAL CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION, ACII, 2023,
  • [28] Vision-Language Pre-Training with Triple Contrastive Learning
    Yang, Jinyu
    Duan, Jiali
    Tran, Son
    Xu, Yi
    Chanda, Sampath
    Chen, Liqun
    Zeng, Belinda
    Chilimbi, Trishul
    Huang, Junzhou
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 15650 - 15659
  • [29] A Comparison between Pre-training and Large-scale Back-translation for Neural Machine Translation
    Huang, Dandan
    Wang, Kun
    Zhang, Yue
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 1718 - 1732
  • [30] Neural Machine Translation Based on XLM-R Cross-lingual Pre-training Language Model
    Wang Q.
    Li M.
    Wu S.
    Wang M.
    Beijing Daxue Xuebao (Ziran Kexue Ban)/Acta Scientiarum Naturalium Universitatis Pekinensis, 2022, 58 (01): : 29 - 36