Sequence-Level Knowledge Distillation for Class-Incremental End-to-End Spoken Language Understanding

被引:1
|
作者
Cappellazzo, Umberto [1 ]
Yang, Muqiao [2 ]
Falavigna, Daniele [3 ]
Brutti, Alessio [3 ]
机构
[1] Univ Trento, Trento, Italy
[2] Carnegie Mellon Univ, Pittsburgh, PA USA
[3] Fdn Bruno Kessler, Trento, Italy
来源
关键词
continual learning; spoken language understanding; knowledge distillation; transformer; NEURAL-NETWORKS;
D O I
10.21437/Interspeech.2023-242
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
The ability to learn new concepts sequentially is a major weakness for modern neural networks, which hinders their use in non-stationary environments. Their propensity to fit the current data distribution to the detriment of the past acquired knowledge leads to the catastrophic forgetting issue. In this work we tackle the problem of Spoken Language Understanding applied to a continual learning setting. We first define a class-incremental scenario for the SLURP dataset. Then, we propose three knowledge distillation (KD) approaches to mitigate forgetting for a sequence-to-sequence transformer model: the first KD method is applied to the encoder output (audio-KD), and the other two work on the decoder output, either directly on the token-level (tok-KD) or on the sequence-level (seq-KD) distributions. We show that the seq-KD substantially improves all the performance metrics, and its combination with the audioKD further decreases the average WER and enhances the entity prediction metric.
引用
收藏
页码:2953 / 2957
页数:5
相关论文
共 50 条
  • [41] END-to-END Cross-Lingual Spoken Language Understanding Model with Multilingual Pretraining
    Zhang, Xianwei
    He, Liang
    INTERSPEECH 2021, 2021, : 4728 - 4732
  • [42] Decomposed Knowledge Distillation for Class-Incremental Semantic Segmentation
    Baek, Donghyeon
    Oh, Youngmin
    Lee, Sanghoon
    Lee, Junghyup
    Ham, Bumsub
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [43] USE OF KERNEL DEEP CONVEX NETWORKS AND END-TO-END LEARNING FOR SPOKEN LANGUAGE UNDERSTANDING
    Deng, Li
    Tur, Gokhan
    He, Xiaodong
    Hakkani-Tur, Dilek
    2012 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY (SLT 2012), 2012, : 210 - 215
  • [44] Investigating Adaptation and Transfer Learning for End-to-End Spoken Language Understanding from Speech
    Tomashenko, Natalia
    Caubriere, Antoine
    Esteve, Yannick
    INTERSPEECH 2019, 2019, : 824 - 828
  • [45] Adapting Transformer to End-to-end Spoken Language Translation
    Di Gangi, Mattia A.
    Negri, Matteo
    Turchi, Marco
    INTERSPEECH 2019, 2019, : 1133 - 1137
  • [46] Diverse Knowledge Distillation for End-to-End Person Search
    Zhang, Xinyu
    Wang, Xinlong
    Bian, Jia-Wang
    Shen, Chunhua
    You, Mingyu
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 3412 - 3420
  • [47] LARGE-SCALE UNSUPERVISED PRE-TRAINING FOR END-TO-END SPOKEN LANGUAGE UNDERSTANDING
    Wang, Pengwei
    Wei, Liangchen
    Cao, Yong
    Xie, Jinghui
    Nie, Zaiqing
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 7999 - 8003
  • [48] DIALOGUE HISTORY INTEGRATION INTO END-TO-END SIGNAL-TO-CONCEPT SPOKEN LANGUAGE UNDERSTANDING SYSTEMS
    Tomashenko, Natalia
    Raymond, Christian
    Caubriere, Antoine
    De Mori, Renato
    Esteve, Yannick
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 8509 - 8513
  • [49] The Interpreter Understands Your Meaning: End-to-end Spoken Language Understanding Aided by Speech Translation
    He, Mutian
    Garner, Philip N.
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 4408 - 4423
  • [50] Class-Incremental Learning by Knowledge Distillation with Adaptive Feature Consolidation
    Kang, Minsoo
    Park, Jaeyoo
    Han, Bohyung
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 16050 - 16059