Harnessing Knowledge Distillation for Enhanced Text-to-Text Translation in Low-Resource Languages

被引:0
|
作者
Ahmed, Manar Ouled [1 ]
Ming, Zuheng [3 ]
Othmani, Alice [2 ,4 ]
机构
[1] Declic AI Res, Riyadh, Saudi Arabia
[2] Deck AI Res, Melbourne, Vic, Australia
[3] Univ Sorbonne Paris Nord, L2TI, Villetaneuse, France
[4] Univ Paris Est, UPEC, LISSI, Vitry Sur Seine, France
来源
关键词
Text-to-text; BART; Low-resource languages;
D O I
10.1007/978-3-031-78014-1_22
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Text-to-text translation is crucial for effective communication and understanding across different languages. In this paper, we present a deep learning-based approach for text-to-text translation. Our method leverages knowledge distillation from a high-performing teacher model, specifically the BART model, to train a smaller and more efficient student model, the mBART model. For that, we minimize the cross-entropy between the model distribution and a learned teacher distribution rather than the observed data, to achieve effective knowledge distillation. Our approach mitigates catastrophic forgetting, especially in low-resource languages, by utilizing the complementary knowledge provided by the teacher model. Extensive experimentation and evaluation demonstrate that our model outperforms state-of-the-art methods, achieving superior BLEU scores on benchmark datasets for French-to-Russian, English-to-Dutch, and Russian-to-Vietnamese translations. An ablation study further shows that the combination of fine-tuning and knowledge distillation enhances the student model's ability to capture linguistic nuances and produce more accurate translations.
引用
收藏
页码:295 / 307
页数:13
相关论文
共 50 条
  • [41] End-to-end Text-to-speech for Low-resource Languages by Cross-Lingual Transfer Learning
    Chen, Yuan-Jui
    Tu, Tao
    Yeh, Cheng-chieh
    Lee, Hung-yi
    INTERSPEECH 2019, 2019, : 2075 - 2079
  • [42] Benchmarking Neural and Statistical Machine Translation on Low-Resource African Languages
    Duh, Kevin
    McNamee, Paul
    Post, Matt
    Thompson, Brian
    PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020), 2020, : 2667 - 2675
  • [43] An Analysis of Massively Multilingual Neural Machine Translation for Low-Resource Languages
    Mueller, Aaron
    Nicolai, Garrett
    McCarthy, Arya D.
    Lewis, Dylan
    Wu, Winston
    Yarowsky, David
    PROCEEDINGS OF THE 12TH INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION (LREC 2020), 2020, : 3710 - 3718
  • [44] Towards a Low-Resource Neural Machine Translation for Indigenous Languages in Canada
    Ngoc Tan Le
    Sadat, Fatiha
    TRAITEMENT AUTOMATIQUE DES LANGUES, 2021, 62 (03): : 39 - 63
  • [45] Neural machine translation for low-resource languages without parallel corpora
    Karakanta, Alina
    Dehdari, Jon
    van Genabith, Josef
    MACHINE TRANSLATION, 2018, 32 (1-2) : 167 - 189
  • [46] Adapting NMT to caption translation in Wikimedia Commons for low-resource languages
    Poncelas, Alberto
    Sarasola, Kepa
    Dowling, Meghan
    Way, Andy
    Labaka, Gorka
    Alegria, Inaki
    PROCESAMIENTO DEL LENGUAJE NATURAL, 2019, (63): : 33 - 40
  • [47] Text is Text, No Matter What: Unifying Text Recognition using Knowledge Distillation
    Bhunia, Ayan Kumar
    Sain, Aneeshan
    Chowdhury, Pinaki Nath
    Song, Yi-Zhe
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 963 - 972
  • [48] Hierarchical Keyword Generation Method for Low-Resource Social Media Text
    Guan, Xinyi
    Long, Shun
    INFORMATION, 2023, 14 (11)
  • [49] Prompt Tuning on Graph-Augmented Low-Resource Text Classification
    Wen, Zhihao
    Fang, Yuan
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (12) : 9080 - 9095
  • [50] Regularisation for Efficient Softmax Parameter Generation in Low-Resource Text Classifiers
    Griesshaber, Daniel
    Maucher, Johannes
    Vu, Ngoc Thang
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 5058 - 5066