INVESTIGATION OF SEQUENCE-LEVEL KNOWLEDGE DISTILLATION METHODS FOR CTC ACOUSTIC MODELS

被引:0
|
作者
Takashima, Ryoichi [1 ,2 ,3 ]
Sheng, Li [1 ]
Kawai, Hisashi [1 ]
机构
[1] Natl Inst Informat & Commun Technol NICT, Koganei, Tokyo, Japan
[2] NICT, Koganei, Tokyo, Japan
[3] Hitachi Ltd, Tokyo, Japan
关键词
Speech recognition; acoustic model; connectionist temporal classification; knowledge distillation;
D O I
暂无
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
This paper presents knowledge distillation ( KD) methods for training connectionist temporal classification ( CTC) acoustic models. In a previous study, we proposed a KD method based on the sequence-level cross-entropy, and showed that the conventional KD method based on the frame-level cross-entropy did not work effectively for CTC acoustic models, whereas the proposed method improved the performance of the models. In this paper, we investigate the implementation of sequence-level KD for CTC models and propose a lattice-based sequence-level KD method. Experiments investigating model compression and the training of a noise-robust model using the Wall Street Journal ( WSJ) and CHiME4 datasets demonstrate that the sequence-level KD methods improve the performance of CTC acoustic models on both two tasks, and show that the lattice-based method can compute the sequence-level KD more efficiently than the N-best-based method proposed in our previous work.
引用
收藏
页码:6156 / 6160
页数:5
相关论文
共 26 条
  • [1] Investigation of Sequence-level Knowledge Distillation Methods for CTC Acoustic Models
    Takashima, Ryoichi
    Sheng, Li
    Kawai, Hisashi
    ICASSP, IEEE International Conference on Acoustics, Speech and Signal Processing - Proceedings, 2019, 2019-May : 6156 - 6160
  • [2] AN INVESTIGATION OF A KNOWLEDGE DISTILLATION METHOD FOR CTC ACOUSTIC MODELS
    Takashima, Ryoichi
    Li, Sheng
    Kawai, Hisashi
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 5809 - 5813
  • [3] f-Divergence Minimization for Sequence-Level Knowledge Distillation
    Wen, Yuqiao
    Li, Zichao
    Du, Wenyu
    Mou, Lili
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 10817 - 10834
  • [4] Mutual-learning sequence-level knowledge distillation for automatic speech recognition
    Li, Zerui
    Ming, Yue
    Yang, Lei
    Xue, Jing-Hao
    NEUROCOMPUTING, 2021, 428 : 259 - 267
  • [5] SEQUENCE-LEVEL KNOWLEDGE DISTILLATION FOR MODEL COMPRESSION OF ATTENTION-BASED SEQUENCE-TO-SEQUENCE SPEECH RECOGNITION
    Mun'im, Raden Mu'az
    Inoue, Nakamasa
    Shinoda, Koichi
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 6151 - 6155
  • [6] Sequence-level Confidence Classifier for ASR Utterance Accuracy and Application to Acoustic Models
    Afshan, Amber
    Kumar, Kshitiz
    Wu, Jian
    INTERSPEECH 2021, 2021, : 4084 - 4088
  • [7] Sequence-Level Knowledge Distillation for Class-Incremental End-to-End Spoken Language Understanding
    Cappellazzo, Umberto
    Yang, Muqiao
    Falavigna, Daniele
    Brutti, Alessio
    INTERSPEECH 2023, 2023, : 2953 - 2957
  • [8] INVESTIGATING SEQUENCE-LEVEL NORMALISATION FOR CTC-LIKE END-TO-END ASR
    Zhao, Zeyu
    Bell, Peter
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 7792 - 7796
  • [9] INVESTIGATION OF LATTICE-FREE MAXIMUM MUTUAL INFORMATION-BASED ACOUSTIC MODELS WITH SEQUENCE-LEVEL KULLBACK-LEIBLER DIVERGENCE
    Kanda, Naoyuki
    Fujita, Yusuke
    Nagamatsu, Kenji
    2017 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU), 2017, : 69 - 76
  • [10] SEQUENCE DISTILLATION FOR PURELY SEQUENCE TRAINED ACOUSTIC MODELS
    Kanda, Naoyuki
    Fujita, Yusuke
    Nagamatsu, Kenji
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 5964 - 5968