End-to-End Spoken Language Understanding: Bootstrapping in Low Resource Scenarios

被引:20
|
作者
Bhosale, Swapnil [1 ]
Sheikh, Imran [1 ]
Dumpala, Sri Harsha [1 ]
Kopparapu, Sunil Kumar [1 ]
机构
[1] TCS Res & Innovat Mumbai, Mumbai, Maharashtra, India
来源
关键词
SLU; intent classification; low resource;
D O I
10.21437/Interspeech.2019-2366
中图分类号
R36 [病理学]; R76 [耳鼻咽喉科学];
学科分类号
100104 ; 100213 ;
摘要
End-to-end Spoken Language Understanding (SLU) systems, without speech-to-text conversion, are more promising in low resource scenarios. They can be more effective when there is not enough labeled data to train reliable speech recognition and language understanding systems, or where running SLU on edge is preferred over cloud based services. In this paper, we present an approach for bootstrapping end-to-end SLU in low resource scenarios. We show that incorporating layers extracted from pre-trained acoustic models, instead of using the typical Mel filter bank features, lead to better performing SLU models. Moreover, the layers extracted from a model pre-trained on one language perform well even for (a) SLU tasks on a different language and also (b) on utterances from speakers with speech disorder.
引用
收藏
页码:1188 / 1192
页数:5
相关论文
共 50 条
  • [31] ATTENTIVE CONTEXTUAL CARRYOVER FOR MULTI-TURN END-TO-END SPOKEN LANGUAGE UNDERSTANDING
    Wei, Kai
    Tran, Thanh
    Chang, Feng-Ju
    Sathyendra, Kanthashree Mysore
    Muniyappa, Thejaswi
    Hu, Jing
    Raju, Anirudh
    McGowan, Ross
    Susanj, Nathan
    Rastrow, Ariya
    Strimel, Grant P.
    2021 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU), 2021, : 837 - 844
  • [32] TWO-STAGE TEXTUAL KNOWLEDGE DISTILLATION FOR END-TO-END SPOKEN LANGUAGE UNDERSTANDING
    Kim, Seongbin
    Kim, Gyuwan
    Shin, Seongjin
    Lee, Sangmin
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7463 - 7467
  • [33] Efficient Adaptation of Spoken Language Understanding based on End-to-End Automatic Speech Recognition
    Kim, Eesung
    Jajodia, Aditya
    Tseng, Cindy
    Neelagiri, Divya
    Ki, Taeyeon
    Apsingekar, Vijendra Raj
    INTERSPEECH 2023, 2023, : 3959 - 3963
  • [34] Rethinking End-to-End Evaluation of Decomposable Tasks: A Case Study on Spoken Language Understanding
    Arora, Siddhant
    Ostapenko, Alissa
    Viswanathan, Vijay
    Dalmia, Siddharth
    Metze, Florian
    Watanabe, Shinji
    Black, Alan W.
    INTERSPEECH 2021, 2021, : 1264 - 1268
  • [35] Speak or Chat with Me: End-to-End Spoken Language Understanding System with Flexible Inputs
    Cha, Sujeong
    Hou, Wangrui
    Jung, Hyun
    Phung, My
    Picheny, Michael
    Kuo, Hong-Kwang J.
    Thomas, Samuel
    Morais, Edmilson
    INTERSPEECH 2021, 2021, : 4723 - 4727
  • [36] END-to-END Cross-Lingual Spoken Language Understanding Model with Multilingual Pretraining
    Zhang, Xianwei
    He, Liang
    INTERSPEECH 2021, 2021, : 4728 - 4732
  • [37] USE OF KERNEL DEEP CONVEX NETWORKS AND END-TO-END LEARNING FOR SPOKEN LANGUAGE UNDERSTANDING
    Deng, Li
    Tur, Gokhan
    He, Xiaodong
    Hakkani-Tur, Dilek
    2012 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY (SLT 2012), 2012, : 210 - 215
  • [38] Investigating Adaptation and Transfer Learning for End-to-End Spoken Language Understanding from Speech
    Tomashenko, Natalia
    Caubriere, Antoine
    Esteve, Yannick
    INTERSPEECH 2019, 2019, : 824 - 828
  • [39] Adapting Transformer to End-to-end Spoken Language Translation
    Di Gangi, Mattia A.
    Negri, Matteo
    Turchi, Marco
    INTERSPEECH 2019, 2019, : 1133 - 1137
  • [40] End-to-End Speech Recognition with Deep Fusion: Leveraging External Language Models for Low-Resource Scenarios
    Zhang, Lusheng
    Wu, Shie
    Wang, Zhongxun
    ELECTRONICS, 2025, 14 (04):