Harnessing Knowledge Distillation for Enhanced Text-to-Text Translation in Low-Resource Languages

被引:0
|
作者
Ahmed, Manar Ouled [1 ]
Ming, Zuheng [3 ]
Othmani, Alice [2 ,4 ]
机构
[1] Declic AI Res, Riyadh, Saudi Arabia
[2] Deck AI Res, Melbourne, Vic, Australia
[3] Univ Sorbonne Paris Nord, L2TI, Villetaneuse, France
[4] Univ Paris Est, UPEC, LISSI, Vitry Sur Seine, France
来源
关键词
Text-to-text; BART; Low-resource languages;
D O I
10.1007/978-3-031-78014-1_22
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Text-to-text translation is crucial for effective communication and understanding across different languages. In this paper, we present a deep learning-based approach for text-to-text translation. Our method leverages knowledge distillation from a high-performing teacher model, specifically the BART model, to train a smaller and more efficient student model, the mBART model. For that, we minimize the cross-entropy between the model distribution and a learned teacher distribution rather than the observed data, to achieve effective knowledge distillation. Our approach mitigates catastrophic forgetting, especially in low-resource languages, by utilizing the complementary knowledge provided by the teacher model. Extensive experimentation and evaluation demonstrate that our model outperforms state-of-the-art methods, achieving superior BLEU scores on benchmark datasets for French-to-Russian, English-to-Dutch, and Russian-to-Vietnamese translations. An ablation study further shows that the combination of fine-tuning and knowledge distillation enhances the student model's ability to capture linguistic nuances and produce more accurate translations.
引用
收藏
页码:295 / 307
页数:13
相关论文
共 50 条
  • [21] A Systematic Review and Analysis of Multilingual Data Strategies in Text-to-Speech for Low-Resource Languages
    Do, Phat
    Coler, Matt
    Dijkstra, Jelske
    Klabbers, Esther
    INTERSPEECH 2021, 2021, : 16 - 20
  • [22] XAlign: Cross-lingual Fact-to-Text Alignment and Generation for Low-Resource Languages
    Abhishek, Tushar
    Sagare, Shivprasad
    Singh, Bhavyajeet
    Sharma, Anubhav
    Gupta, Manish
    Varma, Vasudeva
    COMPANION PROCEEDINGS OF THE WEB CONFERENCE 2022, WWW 2022 COMPANION, 2022, : 171 - 175
  • [23] mT6: Multilingual Pretrained Text-to-Text Transformer with Translation Pairs
    Chi, Zewen
    Dong, Li
    Ma, Shuming
    Huang, Shaohan
    Mao, Xian-Ling
    Huang, Heyan
    Wei, Furu
    2021 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2021), 2021, : 1671 - 1683
  • [24] Hybrid Approach Text Generation for Low-Resource Language
    Rakhimova, Diana
    Adali, Esref
    Karibayeva, Aidana
    ADVANCES IN COMPUTATIONAL COLLECTIVE INTELLIGENCE, ICCCI 2024, PART I, 2024, 2165 : 256 - 268
  • [25] Knowledge Transfer for Utterance Classification in Low-Resource Languages
    Smirnov, Andrei
    Mendelev, Valentin
    SPEECH AND COMPUTER, 2016, 9811 : 435 - 442
  • [26] Introduction to the second issue on machine translation for low-resource languages
    Liu, Chao-Hong
    Karakanta, Alina
    Tong, Audrey N.
    Aulov, Oleg
    Soboroff, Ian M.
    Washington, Jonathan
    Zhao, Xiaobing
    MACHINE TRANSLATION, 2021, 35 (01) : 1 - 2
  • [27] Machine Translation in Low-Resource Languages by an Adversarial Neural Network
    Sun, Mengtao
    Wang, Hao
    Pasquine, Mark
    Hameed, Ibrahim A.
    APPLIED SCIENCES-BASEL, 2021, 11 (22):
  • [28] Research on morphological knowledge-guided low-resource agglutinative languages-Chinese translation
    Abudouwaili, Gulinigeer
    Ruzmamat, Sirajahmat
    Abiderexiti, Kahaerjiang
    Yibulayin, Tuergen
    Yi, Nian
    Wumaier, Aishan
    COMPLEX & INTELLIGENT SYSTEMS, 2025, 11 (03)
  • [29] Neural Machine Translation of Low-Resource and Similar Languages with Backtranslation
    Przystupa, Michael
    Abdul-Mageed, Muhammad
    FOURTH CONFERENCE ON MACHINE TRANSLATION (WMT 2019), VOL 3: SHARED TASK PAPERS, DAY 2, 2019, : 224 - 235
  • [30] Extremely low-resource neural machine translation for Asian languages
    Rubino, Raphael
    Marie, Benjamin
    Dabre, Raj
    Fujita, Atushi
    Utiyama, Masao
    Sumita, Eiichiro
    MACHINE TRANSLATION, 2020, 34 (04) : 347 - 382