A Fast Attention Network for Joint Intent Detection and Slot Filling on Edge Devices

被引:1
|
作者
Huang L. [1 ]
Liang S. [2 ]
Ye F. [2 ]
Gao N. [1 ]
机构
[1] College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou
[2] College of Information Engineering, Zhejiang University of Technology, Hangzhou
来源
基金
中国国家自然科学基金;
关键词
Attention network; edge devices; inference latency; intent detection; natural language understanding (NLU);
D O I
10.1109/TAI.2023.3309272
中图分类号
学科分类号
摘要
Intent detection and slot filling are two main tasks in natural language understanding and play an essential role in task-oriented dialogue systems. The joint learning of both tasks can improve inference accuracy and is popular in recent works. However, most joint models ignore the inference latency and cannot meet the need to deploy dialogue systems at the edge. In this article, we propose a fast attention network (FAN) for joint intent detection and slot filling tasks, guaranteeing both accuracy and latency. Specifically, we introduce a clean and parameter-refined attention module to enhance the information exchange between intent and slot, improving semantic accuracy by more than 2%. The FAN can be implemented on different encoders and delivers more accurate models at every speed level. Our experiments on the Jetson Nano platform show that the FAN inferences 15 utterances per second with a small accuracy drop, showing its effectiveness and efficiency on edge devices. © 2023 IEEE.
引用
收藏
页码:530 / 540
页数:10
相关论文
共 50 条
  • [1] WAIS: Word Attention for Joint Intent Detection and Slot Filling
    Chen, Sixuan
    Yu, Shuai
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 9927 - 9928
  • [2] Attention-Based Recurrent Neural Network Models for Joint Intent Detection and Slot Filling
    Liu, Bing
    Lane, Ian
    17TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2016), VOLS 1-5: UNDERSTANDING SPEECH PROCESSING IN HUMANS AND MACHINES, 2016, : 685 - 689
  • [3] Joint intent detection and slot filling with wheel-graph attention networks
    Wei, Pengfei
    Zeng, Bi
    Liao, Wenxiong
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2022, 42 (03) : 2409 - 2420
  • [4] CEA-Net: a co-interactive external attention network for joint intent detection and slot filling
    Wu D.
    Jiang L.
    Yin L.
    Li Z.
    Huang H.
    Neural Computing and Applications, 2024, 36 (22) : 13513 - 13525
  • [5] Joint agricultural intent detection and slot filling based on enhanced heterogeneous attention mechanism
    Hao, Xia
    Wang, Lu
    Zhu, Hongmei
    Guo, Xuchao
    COMPUTERS AND ELECTRONICS IN AGRICULTURE, 2023, 207
  • [6] SlotRefine: A Fast Non-Autoregressive Model for Joint Intent Detection and Slot Filling
    Wu, Di
    Ding, Liang
    Lu, Fan
    Xie, Jian
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 1932 - 1937
  • [7] CONVOLUTIONAL NEURAL NETWORK BASED TRIANGULAR CRF FOR JOINT INTENT DETECTION AND SLOT FILLING
    Xu, Puyang
    Sarikaya, Ruhi
    2013 IEEE WORKSHOP ON AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING (ASRU), 2013, : 78 - 83
  • [8] Dirichlet variational autoencoder for joint slot filling and intent detection
    Gao, Wang
    Wang, Yu-Wei
    Zhang, Fan
    Fang, Yuan
    Journal of Computers (Taiwan), 2021, 32 (02) : 61 - 73
  • [9] ELSF: Entity-Level Slot Filling Framework for Joint Multiple Intent Detection and Slot Filling
    Zhu, Zhanbiao
    Huang, Peijie
    Huang, Haojing
    Xu, Yuhong
    Lin, Piyuan
    Lao, Leyi
    Chen, Shaoshen
    Xie, Haojie
    Yin, Shangjian
    IEEE/ACM Transactions on Audio Speech and Language Processing, 2024, 32 : 4880 - 4893
  • [10] Intent-Slot Correlation Modeling for Joint Intent Prediction and Slot Filling
    Fan, Jun-Feng
    Wang, Mei-Ling
    Li, Chang-Liang
    Zhu, Zi-Qiang
    Mao, Lu
    JOURNAL OF COMPUTER SCIENCE AND TECHNOLOGY, 2022, 37 (02) : 309 - 319