Joint syntactic and semantic analysis, with a multitask Deep Learning Framework for Spoken Language Understanding

被引:2
|
作者
Tafforeau, Jeremie [1 ]
Bechet, Frederic [1 ]
Artiere, Thierry [1 ,2 ]
Favre, Benoit [1 ]
机构
[1] Aix Marseille Univ, CNRS, LIF, UMR 7279, F-13000 Marseille, France
[2] Ecole Cent Marseille, F-13000 Marseille, France
关键词
Spoken Language Understanding; Recurrent Neural Networks; Long Short Term Memory; FrameNet parsing; Multitask;
D O I
10.21437/Interspeech.2016-851
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Spoken Language Understanding (SLU) models have to deal with Automatic Speech Recognition outputs which are prone to contain errors. Most of SLU models overcome this issue by directly predicting semantic labels from words without any deep linguistic analysis. This is acceptable when enough training data is available to train SLU models in a supervised way. However for open-domain SLU, such annotated corpus is not easily available or very expensive to obtain, and generic syntactic and semantic models, such as dependency parsing, Semantic Role Labeling (SRL) or FrameNet parsing are good candidates if they can be applied to noisy ASR transcriptions with enough robustness. To tackle this issue we present in this paper an RNN-based architecture for performing joint syntactic and semantic parsing tasks on noisy ASR outputs. Experiments carried on a corpus of French spoken conversations collected in a telephone call-centre are reported and show that our strategy brings an improvement over the standard pipeline approach while allowing a lot more flexibility in the model design and optimization.
引用
收藏
页码:3260 / 3264
页数:5
相关论文
共 50 条
  • [31] Deep Stacked Autoencoders for Spoken Language Understanding
    Janod, Killian
    Morchid, Mohamed
    Dufour, Richard
    Linares, Georges
    De Mori, Renato
    [J]. 17TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2016), VOLS 1-5: UNDERSTANDING SPEECH PROCESSING IN HUMANS AND MACHINES, 2016, : 720 - 724
  • [32] A BERT BASED JOINT LEARNING MODEL WITH FEATURE GATED MECHANISM FOR SPOKEN LANGUAGE UNDERSTANDING
    Zhang, Wang
    Jiang, Lei
    Zhang, Shaokang
    Wang, Shuo
    Tan, Jianlong
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 7512 - 7516
  • [33] Pre-training for Spoken Language Understanding with Joint Textual and Phonetic Representation Learning
    Chen, Qian
    Wang, Wen
    Zhang, Qinglin
    [J]. INTERSPEECH 2021, 2021, : 1244 - 1248
  • [34] JOINT GENERATIVE AND DISCRIMINATIVE MODELS FOR SPOKEN LANGUAGE UNDERSTANDING
    Dinarelli, Marco
    Moschitti, Alessandro
    Riccardi, Giuseppe
    [J]. 2008 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY: SLT 2008, PROCEEDINGS, 2008, : 61 - 64
  • [35] STATISTICAL SEMANTIC INTERPRETATION MODELING FOR SPOKEN LANGUAGE UNDERSTANDING WITH ENRICHED SEMANTIC FEATURES
    Celikyilmaz, Asli
    Hakkani-Tuer, Dilek
    Tur, Gokhan
    [J]. 2012 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY (SLT 2012), 2012, : 216 - 221
  • [36] Learning Dialogue History for Spoken Language Understanding
    Zhang, Xiaodong
    Ma, Dehong
    Wang, Houfeng
    [J]. NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, PT I, 2018, 11108 : 120 - 132
  • [37] Learning with noisy supervision for Spoken Language Understanding
    Raymond, Christian
    Riccardi, Giuseppe
    [J]. 2008 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING, VOLS 1-12, 2008, : 4989 - +
  • [38] SPEECH UNDERSTANDING THROUGH SYNTACTIC AND SEMANTIC ANALYSIS
    WALKER, DE
    [J]. IEEE TRANSACTIONS ON COMPUTERS, 1976, 25 (04) : 432 - 439
  • [39] Transfer Learning of Transformers for Spoken Language Understanding
    Svec, Jan
    Fremund, Adam
    Bulin, Martin
    Lehecka, Jan
    [J]. TEXT, SPEECH, AND DIALOGUE (TSD 2022), 2022, 13502 : 489 - 500
  • [40] Transfer Learning Methods for Spoken Language Understanding
    Wang, Xu
    Tang, Chengda
    Zhao, Xiaotian
    Li, Xuancai
    Jin, Zhuolin
    Zheng, Dequan
    Zhao, Tiejun
    [J]. ICMI'19: PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2019, : 510 - 515