SPEECH RECOGNITION BY SIMPLY FINE-TUNING BERT

被引:10
|
作者
Huang, Wen-Chin [1 ,2 ]
Wu, Chia-Hua [2 ]
Luo, Shang-Bao [2 ]
Chen, Kuan-Yu [3 ]
Wang, Hsin-Min [2 ]
Toda, Tomoki [1 ]
机构
[1] Nagoya Univ, Nagoya, Aichi, Japan
[2] Acad Sinica, Taipei, Taiwan
[3] Natl Taiwan Univ Sci & Technol, Taipei, Taiwan
关键词
speech recognition; BERT; language model;
D O I
10.1109/ICASSP39728.2021.9413668
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
We propose a simple method for automatic speech recognition (ASR) by fine-tuning BERT, which is a language model (LM) trained on large-scale unlabeled text data and can generate rich contextual representations. Our assumption is that given a history context sequence, a powerful LM can narrow the range of possible choices and the speech signal can be used as a simple clue. Hence, comparing to conventional ASR systems that train a powerful acoustic model (AM) from scratch, we believe that speech recognition is possible by simply fine-tuning a BERT model. As an initial study, we demonstrate the effectiveness of the proposed idea on the AISHELL dataset and show that stacking a very simple AM on top of BERT can yield reasonable performance.
引用
收藏
页码:7343 / 7347
页数:5
相关论文
共 50 条
  • [1] Fine-Tuning BERT Model for Materials Named Entity Recognition
    Zhao, Xintong
    Greenberg, Jane
    An, Yuan
    Hu, Xiaohua Tony
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2021, : 3717 - 3720
  • [2] A GA-Based Approach to Fine-Tuning BERT for Hate Speech Detection
    Madukwe, Kosisochukwu Judith
    Gao, Xiaoying
    Xue, Bing
    [J]. 2020 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2020, : 2821 - 2828
  • [3] Jointly Fine-Tuning "BERT-like" Self Supervised Models to Improve Multimodal Speech Emotion Recognition
    Siriwardhana, Shamane
    Reis, Andrew
    Weerasekera, Rivindu
    Nanayakkara, Suranga
    [J]. INTERSPEECH 2020, 2020, : 3755 - 3759
  • [4] Knowledge-based BERT word embedding fine-tuning for emotion recognition
    Zhu, Zixiao
    Mao, Kezhi
    [J]. NEUROCOMPUTING, 2023, 552
  • [5] Transfer fine-tuning of BERT with phrasal paraphrases
    Arase, Yuki
    Tsujii, Junichi
    [J]. COMPUTER SPEECH AND LANGUAGE, 2021, 66
  • [6] Efficient Fine-Tuning of BERT Models on the Edge
    Vucetic, Danilo
    Tayaranian, Mohammadreza
    Ziaeefard, Maryam
    Clark, James J.
    Meyer, Brett H.
    Gross, Warren J.
    [J]. 2022 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS 22), 2022, : 1838 - 1842
  • [7] Transfer Fine-Tuning: A BERT Case Study
    Arase, Yuki
    Tsujii, Junichi
    [J]. 2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 5393 - 5404
  • [8] Investigating Learning Dynamics of BERT Fine-Tuning
    Hao, Yaru
    Dong, Li
    Wei, Furu
    Xu, Ke
    [J]. 1ST CONFERENCE OF THE ASIA-PACIFIC CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 10TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (AACL-IJCNLP 2020), 2020, : 87 - 92
  • [9] Chinese Medical Named Entity Recognition based on Expert Knowledge and Fine-tuning Bert
    Zhang, Bofeng
    Yao, Xiuhong
    Li, Haiyan
    Aini, Mirensha
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON KNOWLEDGE GRAPH, ICKG, 2023, : 84 - 90
  • [10] Patent classification by fine-tuning BERT language model
    Lee, Jieh-Sheng
    Hsiang, Jieh
    [J]. WORLD PATENT INFORMATION, 2020, 61