共 40 条
- [1] DEVLIN J, CHANG M W, LEE K, Et al., BERT: pre-training of deep bidirectional transformers for language understanding, (2018)
- [2] SENNRICH R, HADDOW B, BIRCH A., Improving neural machine translation models with monolingual data, (2015)
- [3] PHAM N L, NGUYEN V V, PHAM T V., A data augmentation method for English-Vietnamese neural machine translation, IEEE Access, 11, pp. 28034-28044, (2023)
- [4] LAMAR A, KAYA Z., Measuring the impact of data augmentation methods for extremely low-resource NMT, Proceedings of the Sixth Workshop on Technologies for Machine Translation of Low-Resource Languages (LoResMT 2023), pp. 101-109, (2023)
- [5] CAI D, WANG Y, LI H Y, Et al., Neural machine translation with monolingual translation memory, (2021)
- [6] LIU Y H, GU J T, GOYAL N, Et al., Multilingual denoising pre-training for neural machine translation, (2020)
- [7] VAKHARIA P, VIGNESH S S, BASMATKAR P., Low-resource formality controlled NMT using pre-trained LM, Proceedings of the 20th International Conference on Spoken Language Translation (IWSLT 2023), pp. 321-329, (2023)
- [8] EDUNOV S, OTT M, AULI M, Et al., Understanding back-translation at scale, (2018)
- [9] HASSAN H, AUE A, CHEN C, Et al., Achieving human parity on automatic Chinese to English news translation, (2018)
- [10] NG N, YEE K, BAEVSKI A, Et al., Facebook FAIR's WMT19 news translation task submission, (2019)