SUPERVISED AND UNSUPERVISED ACTIVE LEARNING FOR AUTOMATIC SPEECH RECOGNITION OF LOW-RESOURCE LANGUAGES

被引:0
|
作者
Syed, Ali Raza [1 ]
Rosenberg, Andrew [1 ]
Kislal, Ellen [2 ]
机构
[1] CUNY, Grad Ctr, New York, NY 10017 USA
[2] IBM TJ Watson Res Ctr, Yorktown Hts, NY USA
关键词
supervised active learning; unsupervised active learning; limited-resource automatic speech recognition; active learning;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Automatic speech recognition (ASR) systems rely on large quantities of transcribed acoustic data. The collection of audio data is relatively cheap, whereas the transcription of that data is relatively expensive. Thus there is an interest in the ASR community in active learning, in which only a small subset of highly representative data chosen from a large pool of untranscribed audio need be transcribed in order to approach the performance of the system trained with much larger amounts of transcribed audio. In this paper, we compare two basic approaches to active learning: a supervised approach in which we build a speech recognition system from a small amount of seed data in order to make the selection of a limited amount of additional audio for transcription, and an unsupervised approach in which no intermediate system recognition system built from seed data is necessary. Our best unsupervised approach performs quite close to our supervised approach, with both outperforming a random selection scheme.
引用
收藏
页码:5320 / 5324
页数:5
相关论文
共 50 条
  • [1] Comparison of Unsupervised Learning and Supervised Learning with Noisy Labels for Low-Resource Speech Recognition
    Schraner, Yanick
    Scheller, Christian
    Pluess, Michel
    Neukom, Lukas
    Vogel, Manfred
    [J]. INTERSPEECH 2022, 2022, : 4875 - 4879
  • [2] Improving Automatic Speech Recognition Performance for Low-Resource Languages With Self-Supervised Models
    Zhao, Jing
    Zhang, Wei-Qiang
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2022, 16 (06) : 1227 - 1241
  • [3] A Method Improves Speech Recognition with Contrastive Learning in Low-Resource Languages
    Sun, Lixu
    Yolwas, Nurmemet
    Jiang, Lina
    [J]. APPLIED SCIENCES-BASEL, 2023, 13 (08):
  • [4] AUTOMATIC RATING OF SPONTANEOUS SPEECH FOR LOW-RESOURCE LANGUAGES
    Al-Ghezi, Ragheb
    Getman, Yaroslav
    Voskoboinik, Ekaterina
    Singh, Mittul
    Kurimo, Mikko
    [J]. 2022 IEEE SPOKEN LANGUAGE TECHNOLOGY WORKSHOP, SLT, 2022, : 339 - 345
  • [5] Speech recognition datasets for low-resource Congolese languages
    Kimanuka, Ussen
    Maina, Ciira wa
    Buyuk, Osman
    [J]. DATA IN BRIEF, 2024, 52
  • [6] Unsupervised Learning For Sequence-to-sequence Text-to-speech For Low-resource Languages
    Zhang, Haitong
    Lin, Yue
    [J]. INTERSPEECH 2020, 2020, : 3161 - 3165
  • [7] OpenASR21: The Second Open Challenge for Automatic Speech Recognition of Low-Resource Languages
    Peterson, Kay
    Tong, Audrey
    Yu, Yan
    [J]. INTERSPEECH 2022, 2022, : 4895 - 4899
  • [8] OpenASR20: An Open Challenge for Automatic Speech Recognition of Conversational Telephone Speech in Low-Resource Languages
    Peterson, Kay
    Tong, Audrey
    Yu, Yan
    [J]. INTERSPEECH 2021, 2021, : 4324 - 4328
  • [9] SEMI-SUPERVISED TRANSFER LEARNING FOR LANGUAGE EXPANSION OF END-TO-END SPEECH RECOGNITION MODELS TO LOW-RESOURCE LANGUAGES
    Kim, Jiyeon
    Kumar, Mehul
    Gowda, Dhananjaya
    Garg, Abhinav
    Kim, Chanwoo
    [J]. 2021 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU), 2021, : 984 - 988
  • [10] A Grounded Unsupervised Universal Part-of-Speech Tagger for Low-Resource Languages
    Cardenas, Ronald
    Lin, Ying
    Ji, Heng
    May, Jonathan
    [J]. 2019 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES (NAACL HLT 2019), VOL. 1, 2019, : 2428 - 2439