Capsule Networks for Low Resource Spoken Language Understanding

被引:21
|
作者
Renkens, Vincent [1 ]
Van Hamme, Hugo [1 ]
机构
[1] KULeuven, Dept Elect Engn, ESAT, Kasteelpk Arenberg 10,Bus 2441, B-3001 Leuven, Belgium
关键词
Spoken Language Understanding; Capsule Networks; Deep Learning; Low Resource;
D O I
10.21437/Interspeech.2018-1013
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Designing a spoken language understanding system for command-and-control applications can be challenging because of a wide variety of domains and users or because of a lack of training data. In this paper we discuss a system that learns from scratch from user demonstrations. This method has the advantage that the same system can be used for many domains and users without modifications and that no training data is required prior to deployment. The user is required to train the system, so for a user friendly experience it is crucial to minimize the required amount of data. In this paper we investigate whether a capsule network can make efficient use of the limited amount of available training data. We compare the proposed model to an approach based on Non-negative Matrix Factorisation which is the state-of-the-art in this setting and another deep learning approach that was recently introduced for end-to-end spoken language understanding. We show that the proposed model outperforms the baseline models for three command-and-control applications: controlling a small robot, a vocally guided card game and a home automation task.
引用
收藏
页码:601 / 605
页数:5
相关论文
共 50 条
  • [21] Mining Polysemous Triplets with Recurrent Neural Networks for Spoken Language Understanding
    Vukotic, Vedran
    Raymond, Christian
    INTERSPEECH 2019, 2019, : 1178 - 1182
  • [22] Using Word Confusion Networks for Slot Filling in Spoken Language Understanding
    Yang, Xiaohao
    Liu, Jia
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 1353 - 1357
  • [23] Spoken language understanding software for language learning
    Alam, Hassan
    Kumar, Aman
    Rahman, Fuad
    Hartono, Rachmat
    Tarnikova, Yuliya
    INT CONF ON CYBERNETICS AND INFORMATION TECHNOLOGIES, SYSTEMS AND APPLICATIONS/INT CONF ON COMPUTING, COMMUNICATIONS AND CONTROL TECHNOLOGIES, VOL II, 2007, : 107 - +
  • [24] Bidirectional internal memory gate recurrent neural networks for spoken language understanding
    Morchid, Mohamed
    INTERNATIONAL JOURNAL OF SPEECH TECHNOLOGY, 2022, 25 (01) : 19 - 27
  • [25] Internal Memory Gate for Recurrent Neural Networks with Application to Spoken Language Understanding
    Morchid, Mohamed
    18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION, 2017, : 3316 - 3319
  • [26] Label-dependency coding in Simple Recurrent Networks for Spoken Language Understanding
    Dinarelli, Marco
    Vukotic, Vedran
    Raymond, Christian
    18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION, 2017, : 2491 - 2495
  • [27] Is it time to switch to Word Embedding and Recurrent Neural Networks for Spoken Language Understanding?
    Vukotic, Vedran
    Raymond, Christian
    Gravier, Guillaume
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 130 - 134
  • [28] Bidirectional internal memory gate recurrent neural networks for spoken language understanding
    Mohamed Morchid
    International Journal of Speech Technology, 2022, 25 : 19 - 27
  • [29] Chinese spoken language understanding in SHTQS
    Mao, Jia-Ju
    Guo, Rong
    Lu, Ru-Zhan
    Journal of Harbin Institute of Technology (New Series), 2005, 12 (02) : 225 - 230
  • [30] System Combination for Spoken Language Understanding
    Hahn, Stefan
    Lehnen, Patrick
    Ney, Hermann
    INTERSPEECH 2008: 9TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2008, VOLS 1-5, 2008, : 236 - 239