A Hybrid Neural Network BERT-Cap Based on Pre-Trained Language Model and Capsule Network for User Intent Classification

被引:3
|
作者
Liu, Hai [1 ,2 ]
Liu, Yuanxia [1 ]
Wong, Leung-Pun [3 ]
Lee, Lap-Kei [3 ]
Hao, Tianyong [1 ,4 ]
机构
[1] South China Normal Univ, Sch Comp Sci, Guangzhou 510000, Peoples R China
[2] Guangzhou Key Lab Big Data & Intelligent Educ, Guangzhou 510000, Peoples R China
[3] Open Univ Hong Kong, Sch Sci & Technol, Kowloon, Hong Kong 999077, Peoples R China
[4] South China Normal Univ, Inst Adv Study Educ Dev Guangdong Hong Kong Macao, Guangzhou 510000, Peoples R China
基金
中国国家自然科学基金;
关键词
Signal encoding - Semantics - Speech processing - Text processing - Encoding (symbols);
D O I
10.1155/2020/8858852
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
User intent classification is a vital component of a question-answering system or a task-based dialogue system. In order to understand the goals of users' questions or discourses, the system categorizes user text into a set of pre-defined user intent categories. User questions or discourses are usually short in length and lack sufficient context; thus, it is difficult to extract deep semantic information from these types of text and the accuracy of user intent classification may be affected. To better identify user intents, this paper proposes a BERT-Cap hybrid neural network model with focal loss for user intent classification to capture user intents in dialogue. The model uses multiple transformer encoder blocks to encode user utterances and initializes encoder parameters with a pre-trained BERT. Then, it extracts essential features using a capsule network with dynamic routing after utterances encoding. Experiment results on four publicly available datasets show that our model BERT-Cap achieves a F1 score of 0.967 and an accuracy of 0.967, outperforming a number of baseline methods, indicating its effectiveness in user intent classification.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Skin Lesion Classification Using Pre-Trained DenseNet201 Deep Neural Network
    Jasil, S. P. Godlin
    Ulagamuthalvi, V.
    [J]. ICSPC'21: 2021 3RD INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND COMMUNICATION (ICPSC), 2021, : 393 - 396
  • [32] An efficient brain tumor detection and classification using pre-trained convolutional neural network models
    Rao, K. Nishanth
    Khalaf, Osamah Ibrahim
    Krishnasree, V.
    Kumar, Aruru Sai
    Alsekait, Deema Mohammed
    Priyanka, S. Siva
    Alattas, Ahmed Saleh
    AbdElminaam, Diaa Salama
    [J]. HELIYON, 2024, 10 (17)
  • [33] Scanned ECG Arrhythmia Classification Using a Pre-trained Convolutional Neural Network as a Feature Extractor
    Aldosari, Hanadi
    Coenen, Frans
    Lip, Gregory Y. H.
    Zheng, Yalin
    [J]. ARTIFICIAL INTELLIGENCE XXXIX, AI 2022, 2022, 13652 : 64 - 80
  • [34] On Cognitive Level Classification of Assessment-items Using Pre-trained BERT-based Model
    Dipto, Adnan Saif
    Limon, Md. Mahmudur Rahman
    Tuba, Fatima Tanjum
    Uddin, Md Mohsin
    Khan, M. Saddam Hossain
    Tuhin, Rashedul Amin
    [J]. PROCEEDINGS OF 2023 7TH INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING AND INFORMATION RETRIEVAL, NLPIR 2023, 2023, : 245 - 251
  • [35] CANCN-BERT: A Joint Pre-Trained Language Model for Classical and Modern Chinese
    Ji, Zijing
    Wang, Xin
    Shen, Yuxin
    Rao, Guozheng
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 3112 - 3116
  • [36] Automated Classification of Urinary Cells: Using Convolutional Neural Network Pre-trained on Lung Cells
    Teramoto, Atsushi
    Michiba, Ayano
    Kiriyama, Yuka
    Sakurai, Eiko
    Shiroki, Ryoichi
    Tsukamoto, Tetsuya
    [J]. APPLIED SCIENCES-BASEL, 2023, 13 (03):
  • [37] BERT-siRNA: siRNA target prediction based on BERT pre-trained interpretable model
    Xu, Jiayu
    Xu, Nan
    Xie, Weixin
    Zhao, Chengkui
    Yu, Lei
    Feng, Weixing
    [J]. GENE, 2024, 910
  • [38] A Filter for SAR Image Despeckling Using Pre-Trained Convolutional Neural Network Model
    Pan, Ting
    Peng, Dong
    Yang, Wen
    Li, Heng-Chao
    [J]. REMOTE SENSING, 2019, 11 (20)
  • [39] Pre-trained Deep Convolution Neural Network Model With Attention for Speech Emotion Recognition
    Zhang, Hua
    Gou, Ruoyun
    Shang, Jili
    Shen, Fangyao
    Wu, Yifan
    Dai, Guojun
    [J]. FRONTIERS IN PHYSIOLOGY, 2021, 12
  • [40] Exponential Discretization of Weights of Neural Network Connections in Pre-Trained Neural Networks
    Malsagov, M. Yu
    Khayrov, E. M.
    Pushkareva, M. M.
    Karandashev, I. M.
    [J]. OPTICAL MEMORY AND NEURAL NETWORKS, 2019, 28 (04) : 262 - 270