LEARNING FROM THE BEST: A TEACHER-STUDENT MULTILINGUAL FRAMEWORK FOR LOW-RESOURCE LANGUAGES

被引:0
|
作者
Bagchi, Deblin [1 ,2 ]
Hartmann, William [2 ]
机构
[1] Ohio State Univ, Dept Comp Sci & Engn, Columbus, OH 43210 USA
[2] Raytheon BBN Technol, Cambridge, MA USA
关键词
Teacher-student learning; Low-resource speech; Multilingual training; Automatic speech recognition;
D O I
10.1109/icassp.2019.8683491
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
The traditional method of pretraining neural acoustic models in low-resource languages consists of initializing the acoustic model parameters with a large, annotated multilingual corpus and can be a drain on time and resources. In an attempt to reuse TDNN-LSTMs already pre-trained using multilingual training, we have applied Teacher-Student ( TS) learning as a method of pretraining to transfer knowledge from a multilingual TDNN-LSTM to a TDNN. The pretraining time is reduced by an order of magnitude with the use of language-specific data during the teacher-student training. Additionally, the TS architecture allows us to leverage untranscribed data, previously untouched during supervised training. The best student TDNN achieves a WER within 1% of the teacher TDNN-LSTM performance and shows consistent improvement in recognition over TDNNs trained using the traditional pipeline over all the evaluation languages. Switching to TDNN from TDNN-LSTM also allows sub-real time decoding.
引用
收藏
页码:6051 / 6055
页数:5
相关论文
共 50 条
  • [21] MuCoT: Multilingual Contrastive Training for Question-Answering in Low-resource Languages
    Kumar, Gokul Karthik
    Gehlot, Abhishek Singh
    Mullappilly, Sahal Shaji
    Nandakumar, Karthik
    PROCEEDINGS OF THE SECOND WORKSHOP ON SPEECH AND LANGUAGE TECHNOLOGIES FOR DRAVIDIAN LANGUAGES (DRAVIDIANLANGTECH 2022), 2022, : 15 - 24
  • [22] Efficient neural speech synthesis for low-resource languages through multilingual modeling
    de Korte, Marcel
    Kim, Jaebok
    Klabbers, Esther
    INTERSPEECH 2020, 2020, : 2967 - 2971
  • [23] Multilingual Contextual Adapters To Improve Custom Word Recognition In Low-resource Languages
    Kulshreshtha, Devang
    Dingliwal, Saket
    Houston, Brady
    Bodapati, Sravan
    INTERSPEECH 2023, 2023, : 3302 - 3306
  • [24] Weighted Cross-entropy for Low-Resource Languages in Multilingual Speech Recognition
    Pineiro-Martin, Andres
    Garcia-Mateo, Carmen
    Docio-Fernandez, Laura
    Del Carmen Lopez-Perez, Maria
    Rehm, Georg
    INTERSPEECH 2024, 2024, : 1235 - 1239
  • [25] A Framework for Motivating Teacher-Student Relationships
    Robinson, Carly D.
    EDUCATIONAL PSYCHOLOGY REVIEW, 2022, 34 (04) : 2061 - 2094
  • [26] A Framework for Motivating Teacher-Student Relationships
    Carly D. Robinson
    Educational Psychology Review, 2022, 34 : 2061 - 2094
  • [27] A self-learning teacher-student framework for gastrointestinal image classification
    Gjestang, Henrik L.
    Hicks, Steven A.
    Thambawita, Vajira
    Halvorsen, Pal
    Riegler, Michael A.
    2021 IEEE 34TH INTERNATIONAL SYMPOSIUM ON COMPUTER-BASED MEDICAL SYSTEMS (CBMS), 2021, : 539 - 544
  • [28] Fixing MoE Over-Fitting on Low-Resource Languages in Multilingual Machine Translation
    Elbayad, Maha
    Sun, Anna
    Bhosale, Shruti
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 14237 - 14253
  • [29] Multilingual end-to-end ASR for low-resource Turkic languages with common alphabets
    Bekarystankyzy, Akbayan
    Mamyrbayev, Orken
    Mendes, Mateus
    Fazylzhanova, Anar
    Assam, Muhammad
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [30] A Transformer-Based Approach to Multilingual Fake News Detection in Low-Resource Languages
    De, Arkadipta
    Bandyopadhyay, Dibyanayan
    Gain, Baban
    Ekbal, Asif
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2022, 21 (01)