DCCL: Dual-channel hybrid neural network combined with self-attention for text classification

被引:0
|
作者
Li, Chaofan [1 ,2 ]
Qiong, Liu [3 ]
Kai, Ma [4 ]
机构
[1] Nanjing Med Univ, Yancheng Sch Clin Med, Nanjing 224008, Jiangsu, Peoples R China
[2] Yancheng Third Peoples Hosp, Qual Management Div, Yancheng 224008, Jiangsu, Peoples R China
[3] Jiangsu Vocat Coll Med, Sch Med Imaging, Yancheng 224005, Jiangsu, Peoples R China
[4] Xuzhou Med Univ, Sch Med Informat & Engn, Xuzhou 221004, Jiangsu, Peoples R China
关键词
text classification; convolutional neural networks; long short-term memory networks; LSTM;
D O I
10.3934/mbe.2023091
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Text classification is a fundamental task in natural language processing. The Chinese text classification task suffers from sparse text features, ambiguity in word segmentation, and poor performance of classification models. A text classification model is proposed based on the self -attention mechanism combined with CNN and LSTM. The proposed model uses word vectors as input to a dual-channel neural network structure, using multiple CNNs to extract the N-Gram information of different word windows and enrich the local feature representation through the concatenation operation, the BiLSTM is used to extract the semantic association information of the context to obtain the high-level feature representation at the sentence level. The output of BiLSTM is feature weighted with self-attention to reduce the influence of noisy features. The outputs of the dual channels are concatenated and fed into the softmax layer for classification. The results of the multiple comparison experiments showed that the DCCL model obtained 90.07% and 96.26% F1-score on the Sougou and THUNews datasets, respectively. Compared to the baseline model, the improvement was 3.24% and 2.19%, respectively. The proposed DCCL model can alleviate the problem of CNN losing word order information and the gradient of BiLSTM when processing text sequences, effectively integrate local and global text features, and highlight key information. The classification performance of the DCCL model is excellent and suitable for text classification tasks.
引用
下载
收藏
页码:1981 / 1992
页数:12
相关论文
共 50 条
  • [21] Multi-Scale Self-Attention for Text Classification
    Guo, Qipeng
    Qiu, Xipeng
    Liu, Pengfei
    Xue, Xiangyang
    Zhang, Zheng
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 7847 - 7854
  • [22] Combining Contextual Information by Self-attention Mechanism in Convolutional Neural Networks for Text Classification
    Wu, Xin
    Cai, Yi
    Li, Qing
    Xu, Jingyun
    Leung, Ho-fung
    WEB INFORMATION SYSTEMS ENGINEERING, WISE 2018, PT I, 2018, 11233 : 453 - 467
  • [23] SAFSN: A Self-Attention Based Neural Network for Encrypted Mobile Traffic Classification
    Zhang, Chengyuan
    An, Changqing
    Wang, Jessie Hui
    Zhao, Ziyi
    Yu, Tao
    Wang, Jilong
    IEEE CONGRESS ON CYBERMATICS / 2021 IEEE INTERNATIONAL CONFERENCES ON INTERNET OF THINGS (ITHINGS) / IEEE GREEN COMPUTING AND COMMUNICATIONS (GREENCOM) / IEEE CYBER, PHYSICAL AND SOCIAL COMPUTING (CPSCOM) / IEEE SMART DATA (SMARTDATA), 2021, : 330 - 337
  • [24] Hyperspectral Image Classification Based on Dual-Channel Dilated Convolution Neural Network
    Hu Li
    Shan Rui
    Wang Fang
    Jiang Guoqian
    Zhao Jingyi
    Zhang Zhi
    LASER & OPTOELECTRONICS PROGRESS, 2020, 57 (12)
  • [25] Classification of Bone Marrow Cells Based on Dual-Channel Convolutional Block Attention Network
    Wang, Zhaorong
    Zheng, Rui
    Zhu, Xiayin
    Luo, Wenda
    He, Sailing
    IEEE ACCESS, 2024, 12 : 96205 - 96219
  • [26] An attention involved network stacked by dual-channel residual block for hyperspectral image classification
    Deng, Ziqing
    Wang, Yang
    Li, Linwei
    Zhang, Bing
    Zhang, Zhengli
    Bian, Lifeng
    Ding, Zhao
    Yang, Chen
    INFRARED PHYSICS & TECHNOLOGY, 2022, 122
  • [27] Dual-channel feature extraction hybrid attention network for detecting infrared small targets
    Nie, Suzhen
    Cao, Jie
    Miao, Jiaqi
    Hou, Haiyuan
    Hao, Qun
    Zhuang, Xuye
    MEASUREMENT SCIENCE AND TECHNOLOGY, 2024, 35 (12)
  • [28] A Dual-channel Attention Model for Optical Microscope Pollen Classification
    Li, Yanan
    Li, Jianqiang
    Pei, Yan
    Wang, Jin
    2022 8TH INTERNATIONAL CONFERENCE ON CONTROL, DECISION AND INFORMATION TECHNOLOGIES (CODIT'22), 2022, : 1118 - 1123
  • [29] TESANet: Self-attention network for olfactory EEG classification
    Tong, Chengxuan
    Ding, Yi
    Liang, Kevin Lim Jun
    Zhang, Zhuo
    Zhang, Haihong
    Guan, Cuntai
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [30] A Dual Self-Attention based Network for Image Captioning
    Li, ZhiYong
    Yang, JinFu
    Li, YaPing
    PROCEEDINGS OF THE 33RD CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2021), 2021, : 1590 - 1595