Harnessing Knowledge Distillation for Enhanced Text-to-Text Translation in Low-Resource Languages

被引:0
|
作者
Ahmed, Manar Ouled [1 ]
Ming, Zuheng [3 ]
Othmani, Alice [2 ,4 ]
机构
[1] Declic AI Res, Riyadh, Saudi Arabia
[2] Deck AI Res, Melbourne, Vic, Australia
[3] Univ Sorbonne Paris Nord, L2TI, Villetaneuse, France
[4] Univ Paris Est, UPEC, LISSI, Vitry Sur Seine, France
来源
关键词
Text-to-text; BART; Low-resource languages;
D O I
10.1007/978-3-031-78014-1_22
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Text-to-text translation is crucial for effective communication and understanding across different languages. In this paper, we present a deep learning-based approach for text-to-text translation. Our method leverages knowledge distillation from a high-performing teacher model, specifically the BART model, to train a smaller and more efficient student model, the mBART model. For that, we minimize the cross-entropy between the model distribution and a learned teacher distribution rather than the observed data, to achieve effective knowledge distillation. Our approach mitigates catastrophic forgetting, especially in low-resource languages, by utilizing the complementary knowledge provided by the teacher model. Extensive experimentation and evaluation demonstrate that our model outperforms state-of-the-art methods, achieving superior BLEU scores on benchmark datasets for French-to-Russian, English-to-Dutch, and Russian-to-Vietnamese translations. An ablation study further shows that the combination of fine-tuning and knowledge distillation enhances the student model's ability to capture linguistic nuances and produce more accurate translations.
引用
收藏
页码:295 / 307
页数:13
相关论文
共 50 条
  • [31] Introduction to the Special Issue on Machine Translation for Low-Resource Languages
    Liu, Chao-Hong
    Karakanta, Alina
    Tong, Audrey N.
    Aulov, Oleg
    Soboroff, Ian M.
    Washington, Jonathan
    Zhao, Xiaobing
    MACHINE TRANSLATION, 2020, 34 (04) : 247 - 249
  • [32] The Impact of Translating Resource-Rich Datasets to Low-Resource Languages Through Multi-Lingual Text Processing
    Ghafoor, Abdul
    Imran, Ali Shariq
    Daudpota, Sher Muhammad
    Kastrati, Zenun
    Abdullah
    Batra, Rakhi
    Wani, Mudasir Ahmad
    IEEE ACCESS, 2021, 9 : 124478 - 124490
  • [33] DISTRIBUTION AUGMENTATION FOR LOW-RESOURCE EXPRESSIVE TEXT-TO-SPEECH
    Lajszczak, Mateusz
    Prasad, Animesh
    van Korlaar, Arent
    Bollepalli, Bajibabu
    Bonafonte, Antonio
    Joly, Arnaud
    Nicolis, Marco
    Moinet, Alexis
    Drugman, Thomas
    Wood, Trevor
    Sokolova, Elena
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 8307 - 8311
  • [34] Prompt-based for Low-Resource Tibetan Text Classification
    An, Bo
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (08)
  • [35] Text Augmentation Using Dataset Reconstruction for Low-Resource Classification
    Rahamim, Adir
    Uziel, Guy
    Goldbraich, Esther
    Anaby-Tavor, Ateret
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 7389 - 7402
  • [36] ParaSum: Contrastive Paraphrasing for Low-Resource Extractive Text Summarization
    Tang, Moming
    Wang, Chengyu
    Wang, Jianing
    Chen, Cen
    Gao, Ming
    Qian, Weining
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT III, KSEM 2023, 2023, 14119 : 106 - 119
  • [37] Enhanced Text Classification using Proxy Labels and Knowledge Distillation
    Sukumaran, Rohan
    Prabhu, Sumanth
    Misra, Hemant
    PROCEEDINGS OF THE 5TH JOINT INTERNATIONAL CONFERENCE ON DATA SCIENCE & MANAGEMENT OF DATA, CODS COMAD 2022, 2022, : 227 - 230
  • [38] Pre-training on High-Resource Speech Recognition Improves Low-Resource Speech-to-Text Translation
    Bansal, Sameer
    Kamper, Herman
    Livescu, Karen
    Lopez, Adam
    Goldwater, Sharon
    2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 58 - 68
  • [39] Resource-Efficient Fine-Tuning Strategies for Automatic MOS Prediction in Text-to-Speech for Low-Resource Languages
    Do, Phat
    Coler, Matt
    Dijkstra, Jelske
    Klabbers, Esther
    INTERSPEECH 2023, 2023, : 5466 - 5470
  • [40] Efficient Neural Machine Translation for Low-Resource Languages via Exploiting Related Languages
    Goyal, Vikrant
    Kumar, Sourav
    Sharma, Dipti Misra
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020): STUDENT RESEARCH WORKSHOP, 2020, : 162 - 168