Deep Stacked Autoencoders for Spoken Language Understanding

被引:3
|
作者
Janod, Killian [1 ,2 ]
Morchid, Mohamed [1 ]
Dufour, Richard [1 ]
Linares, Georges [1 ]
De Mori, Renato [1 ,3 ]
机构
[1] Univ Avignon, LIA, Avignon, France
[2] ORKIS, Aix En Provence, France
[3] McGill Univ, Montreal, PQ, Canada
关键词
spoken dialogues; deep neural networks; denoising autoencoders; deep stacked autoencoders;
D O I
10.21437/Interspeech.2016-63
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
The automatic transcription process of spoken document results in several word errors, especially when very noisy conditions are encountered. Document representations based on neural embedding frameworks have recently shown significant improvements in different Spoken and Natural Language Understanding tasks such as denoising and filtering. Nonetheless, these methods mainly need clean representations, failing to properly remove noise contained in noisy representations. This paper proposes to study the impact of residual noise contained into automatic transcripts of spoken dialogues in highly abstract spaces from deep neural networks. The paper makes the assumption that the noise learned from "clean" manual transcripts of spoken documents moves down dramatically the performance of theme identification systems in noisy conditions. The proposed deep neural network takes, as input and output, highly imperfect transcripts from spoken dialogues to improve the robustness of the document representation in a noisy environment. Results obtained on the DECODA theme classification task of dialogues reach an accuracy of 82% with a significant gain of about 5%.
引用
收藏
页码:720 / 724
页数:5
相关论文
共 50 条
  • [1] Deep Contextual Language Understanding in Spoken Dialogue Systems
    Liu, Chunxi
    Xu, Puyang
    Sarikaya, Ruhi
    [J]. 16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 120 - 124
  • [2] DEEP QUATERNION NEURAL NETWORKS FOR SPOKEN LANGUAGE UNDERSTANDING
    Parcollet, Titouan
    Morchid, Mohamed
    Linares, Georges
    [J]. 2017 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU), 2017, : 504 - 511
  • [3] Research on Spoken Language Understanding Based on Deep Learning
    Yanli Hui
    [J]. SCIENTIFIC PROGRAMMING, 2021, 2021
  • [4] Deep Belief Network based CRF for Spoken Language Understanding
    Yang, Xiaohao
    Liu, Jia
    [J]. 2014 9TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING (ISCSLP), 2014, : 49 - 53
  • [5] Spoken language understanding
    Wang, YY
    Deng, L
    Acero, A
    [J]. IEEE SIGNAL PROCESSING MAGAZINE, 2005, 22 (05) : 16 - 31
  • [6] UNDERSTANDING SPOKEN LANGUAGE
    BROWN, G
    [J]. TESOL QUARTERLY, 1978, 12 (03) : 271 - 283
  • [7] UNSUPERVISED DEEP HASHING WITH STACKED CONVOLUTIONAL AUTOENCODERS
    En, Sovann
    Cremilleux, Bruno
    Jurie, Frederic
    [J]. 2017 24TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2017, : 3420 - 3424
  • [8] Deep Belief Network based Semantic Taggers for Spoken Language Understanding
    Deoras, Anoop
    Sarikaya, Ruhi
    [J]. 14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5, 2013, : 2712 - 2716
  • [9] Intent Detection for Spoken Language Understanding Using a Deep Ensemble Model
    Firdaus, Mauajama
    Bhatnagar, Shobhit
    Ekbal, Asif
    Bhattacharyya, Pushpak
    [J]. PRICAI 2018: TRENDS IN ARTIFICIAL INTELLIGENCE, PT I, 2018, 11012 : 629 - 642
  • [10] Hyperspectral Unmixing Based on Deep Stacked Autoencoders Network
    Zhu Ling
    Qin Kai
    Sun Yu
    Li Ming
    Zhao Ying-jun
    [J]. SPECTROSCOPY AND SPECTRAL ANALYSIS, 2023, 43 (05) : 1508 - 1516