Sequence-Level Knowledge Distillation for Class-Incremental End-to-End Spoken Language Understanding

被引:0
|
作者
Cappellazzo, Umberto [1 ]
Yang, Muqiao [2 ]
Falavigna, Daniele [3 ]
Brutti, Alessio [3 ]
机构
[1] Univ Trento, Trento, Italy
[2] Carnegie Mellon Univ, Pittsburgh, PA USA
[3] Fdn Bruno Kessler, Trento, Italy
来源
关键词
continual learning; spoken language understanding; knowledge distillation; transformer; NEURAL-NETWORKS;
D O I
10.21437/Interspeech.2023-242
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
The ability to learn new concepts sequentially is a major weakness for modern neural networks, which hinders their use in non-stationary environments. Their propensity to fit the current data distribution to the detriment of the past acquired knowledge leads to the catastrophic forgetting issue. In this work we tackle the problem of Spoken Language Understanding applied to a continual learning setting. We first define a class-incremental scenario for the SLURP dataset. Then, we propose three knowledge distillation (KD) approaches to mitigate forgetting for a sequence-to-sequence transformer model: the first KD method is applied to the encoder output (audio-KD), and the other two work on the decoder output, either directly on the token-level (tok-KD) or on the sequence-level (seq-KD) distributions. We show that the seq-KD substantially improves all the performance metrics, and its combination with the audioKD further decreases the average WER and enhances the entity prediction metric.
引用
收藏
页码:2953 / 2957
页数:5
相关论文
共 50 条
  • [31] Confidence measure for speech-to-concept end-to-end spoken language understanding
    Caubriere, Antoine
    Esteve, Yannick
    Laurent, Antoine
    Morin, Emmanuel
    INTERSPEECH 2020, 2020, : 1590 - 1594
  • [32] Speech Model Pre-training for End-to-End Spoken Language Understanding
    Lugosch, Loren
    Ravanelli, Mirco
    Ignoto, Patrick
    Tomar, Vikrant Singh
    Bengio, Yoshua
    INTERSPEECH 2019, 2019, : 814 - 818
  • [33] TOWARDS END-TO-END INTEGRATION OF DIALOG HISTORY FOR IMPROVED SPOKEN LANGUAGE UNDERSTANDING
    Sunder, Vishal
    Thomas, Samuel
    Kuo, Hong-Kwang J.
    Ganhotra, Jatin
    Kingsbury, Brian
    Fosler-Lussier, Eric
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 7497 - 7501
  • [34] SEQUENCE-LEVEL CONSISTENCY TRAINING FOR SEMI-SUPERVISED END-TO-END AUTOMATIC SPEECH RECOGNITION
    Masumura, Ryo
    Ihori, Mana
    Takashima, Akihiko
    Moriya, Takafumi
    Ando, Atsushi
    Shinohara, Yusuke
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 7054 - 7058
  • [35] END-TO-END SPOKEN LANGUAGE UNDERSTANDING WITHOUT MATCHED LANGUAGE SPEECH MODEL PRETRAINING DATA
    Price, Ryan
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 7979 - 7983
  • [36] End-to-End Speech Translation with Knowledge Distillation
    Liu, Yuchen
    Xiong, Hao
    Zhang, Jiajun
    He, Zhongjun
    Wu, Hua
    Wang, Haifeng
    Zong, Chengqing
    INTERSPEECH 2019, 2019, : 1128 - 1132
  • [37] ATTENTIVE CONTEXTUAL CARRYOVER FOR MULTI-TURN END-TO-END SPOKEN LANGUAGE UNDERSTANDING
    Wei, Kai
    Tran, Thanh
    Chang, Feng-Ju
    Sathyendra, Kanthashree Mysore
    Muniyappa, Thejaswi
    Hu, Jing
    Raju, Anirudh
    McGowan, Ross
    Susanj, Nathan
    Rastrow, Ariya
    Strimel, Grant P.
    2021 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU), 2021, : 837 - 844
  • [38] Efficient Adaptation of Spoken Language Understanding based on End-to-End Automatic Speech Recognition
    Kim, Eesung
    Jajodia, Aditya
    Tseng, Cindy
    Neelagiri, Divya
    Ki, Taeyeon
    Apsingekar, Vijendra Raj
    INTERSPEECH 2023, 2023, : 3959 - 3963
  • [39] Rethinking End-to-End Evaluation of Decomposable Tasks: A Case Study on Spoken Language Understanding
    Arora, Siddhant
    Ostapenko, Alissa
    Viswanathan, Vijay
    Dalmia, Siddharth
    Metze, Florian
    Watanabe, Shinji
    Black, Alan W.
    INTERSPEECH 2021, 2021, : 1264 - 1268
  • [40] Speak or Chat with Me: End-to-End Spoken Language Understanding System with Flexible Inputs
    Cha, Sujeong
    Hou, Wangrui
    Jung, Hyun
    Phung, My
    Picheny, Michael
    Kuo, Hong-Kwang J.
    Thomas, Samuel
    Morais, Edmilson
    INTERSPEECH 2021, 2021, : 4723 - 4727