Encoding Syntactic Knowledge in Transformer Encoder for Intent Detection and Slot Filling

被引:0
|
作者
Wang, Jixuan [1 ,2 ,3 ]
Wei, Kai [3 ]
Radfar, Martin [3 ]
Zhang, Weiwei [3 ]
Chung, Clement [3 ]
机构
[1] Univ Toronto, Toronto, ON, Canada
[2] Vector Inst, Toronto, ON, Canada
[3] Amazon Alexa, Pittsburgh, PA 15205 USA
关键词
NEURAL-NETWORKS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a novel Transformer encoder-based architecture with syntactical knowledge encoded for intent detection and slot filling. Specifically, we encode syntactic knowledge into the Transformer encoder by jointly training it to predict syntactic parse ancestors and part-of-speech of each token via multi-task learning. Our model is based on self-attention and feed-forward layers and does not require external syntactic information to be available at inference time. Experiments show that on two benchmark datasets, our models with only two Transformer encoder layers achieve state-of-the-art results. Compared to the previously best performed model without pre-training, our models achieve absolute F1 score and accuracy improvement of 1.59% and 0.85% for slot filling and intent detection on the SNIPS dataset, respectively. Our models also achieve absolute F1 score and accuracy improvement of 0.1% and 0.34% for slot filling and intent detection on the ATIS dataset, respectively, over the previously best performed model. Furthermore, the visualization of the self-attention weights illustrates the benefits of incorporating syntactic information during training.
引用
收藏
页码:13943 / 13951
页数:9
相关论文
共 50 条
  • [21] A Multi-Task Hierarchical Approach for Intent Detection and Slot Filling
    Firdaus, Mauajama
    Kumar, Ankit
    Ekbal, Asif
    Bhattacharyya, Pushpak
    KNOWLEDGE-BASED SYSTEMS, 2019, 183
  • [22] A Graph-to-Sequence Model for Joint Intent Detection and Slot Filling
    Wu, Jie
    Harris, Ian G.
    Zhao, Hongzhi
    Ling, Guangming
    2023 IEEE 17TH INTERNATIONAL CONFERENCE ON SEMANTIC COMPUTING, ICSC, 2023, : 131 - 138
  • [23] Multitask learning for multilingual intent detection and slot filling in dialogue systems
    Firdaus, Mauajama
    Ekbal, Asif
    Cambria, Erik
    INFORMATION FUSION, 2023, 91 : 299 - 315
  • [24] A Deep Learning Model with Data Enrichment for Intent Detection and Slot Filling
    Dadas, Slawomir
    Protasiewicz, Jaroslaw
    Pedrycz, Witold
    2019 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS (SMC), 2019, : 3012 - 3018
  • [25] Joint intent detection and slot filling for Turkish natural language understanding
    Buyuk, Osman
    TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, 2023, 31 (05) : 844 - 859
  • [26] Task Conditioned BERT for Joint Intent Detection and Slot-Filling
    Tavares, Diogo
    Azevedo, Pedro
    Semedo, David
    Sousa, Ricardo
    Magalhaes, Joao
    PROGRESS IN ARTIFICIAL INTELLIGENCE, EPIA 2023, PT I, 2023, 14115 : 467 - 480
  • [27] Joint Slot Filling and Intent Detection via Capsule Neural Networks
    Zhang, Chenwei
    Li, Yaliang
    Du, Nan
    Fan, Wei
    Yu, Philip S.
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 5259 - 5267
  • [28] Decoupling Representation and Knowledge for Few-Shot Intent Classification and Slot Filling
    Han, Jie
    Zou, Yixiong
    Wang, Haozhao
    Wang, Jun
    Liu, Wei
    Wu, Yao
    Zhang, Tao
    Li, Ruixuan
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 16, 2024, : 18171 - 18179
  • [29] JOINT MULTIPLE INTENT DETECTION AND SLOT FILLING VIA SELF-DISTILLATION
    Chen, Lisong
    Zhou, Peilin
    Zou, Yuexian
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 7612 - 7616
  • [30] Joint Intent Detection and Slot Filling via CNN-LSTM-CRF
    Kane, Bamba
    Rossi, Fabio
    Guinaudeau, Ophelie
    Chiesa, Valeria
    Quenel, Ilhem
    Chau, Stephan
    2020 6TH IEEE CONGRESS ON INFORMATION SCIENCE AND TECHNOLOGY (IEEE CIST'20), 2020, : 342 - 347