A Convolutional Attention Model for Text Classification

被引:31
|
作者
Du, Jiachen [1 ,2 ]
Gui, Lin [1 ]
Xu, Ruifeng [1 ,3 ]
He, Yulan [4 ]
机构
[1] Harbin Inst Technol, Shenzhen Grad Sch, Lab Network Oriented Intelligent Computat, Shenzhen, Peoples R China
[2] Hong Kong Polytech Univ, Dept Comp, Hung Hom, Hong Kong, Peoples R China
[3] Guangdong Prov Engn Technol Res Ctr Data Sci, Guangzhou, Guangdong, Peoples R China
[4] Aston Univ, Sch Engn & Appl Sci, Birmingham, W Midlands, England
基金
中国国家自然科学基金;
关键词
D O I
10.1007/978-3-319-73618-1_16
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Neural network models with attention mechanism have shown their efficiencies on various tasks. However, there is little research work on attention mechanism for text classification and existing attention model for text classification lacks of cognitive intuition and mathematical explanation. In this paper, we propose a new architecture of neural network based on the attention model for text classification. In particular, we show that the convolutional neural network (CNN) is a reasonable model for extracting attentions from text sequences in mathematics. We then propose a novel attention model base on CNN and introduce a new network architecture which combines recurrent neural network with our CNN-based attention model. Experimental results on five datasets show that our proposed models can accurately capture the salient parts of sentences to improve the performance of text classification.
引用
收藏
页码:183 / 195
页数:13
相关论文
共 50 条
  • [1] Text Classification Based on Convolutional Neural Network and Attention Model
    Yang, Shuang
    Tang, Yan
    [J]. 2020 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND BIG DATA (ICAIBD 2020), 2020, : 67 - 73
  • [2] Hierarchical Convolutional Attention Networks for Text Classification
    Gao, Shang
    Ramanathan, Arvind
    Tourassi, Georgia
    [J]. REPRESENTATION LEARNING FOR NLP, 2018, : 11 - 23
  • [3] Bidirectional LSTM with attention mechanism and convolutional layer for text classification
    Liu, Gang
    Guo, Jiabao
    [J]. NEUROCOMPUTING, 2019, 337 : 325 - 338
  • [4] ICAN: Introspective Convolutional Attention Network for Semantic Text Classification
    Mondal, Sounak
    Modi, Suraj
    Garg, Sakshi
    Das, Dhruva
    Mukherjee, Siddhartha
    [J]. 2020 IEEE 14TH INTERNATIONAL CONFERENCE ON SEMANTIC COMPUTING (ICSC 2020), 2020, : 158 - 161
  • [5] A Hybrid Bidirectional Recurrent Convolutional Neural Network Attention-Based Model for Text Classification
    Zheng, Jin
    Zheng, Limin
    [J]. IEEE ACCESS, 2019, 7 : 106673 - 106685
  • [6] Audioset classification with Graph Convolutional Attention model
    Li, Xuliang
    Gao, Junbin
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [7] TABAS: Text augmentation based on attention score for text classification model
    Yu, Yeong Jae
    Yoon, Seung Joo
    Jun, So Young
    Kim, Jong Woo
    [J]. ICT EXPRESS, 2022, 8 (04): : 549 - 554
  • [8] Chinese Short Text Classification with Mutual-Attention Convolutional Neural Networks
    Hao, Ming
    Xu, Bo
    Liang, Jing-Yi
    Zhang, Bo-Wen
    Yin, Xu-Cheng
    [J]. ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2020, 19 (05)
  • [9] A Multi-scale Convolutional Attention Based GRU Network for Text Classification
    Tang, Xianlun
    Chen, Yingjie
    Dai, Yuyan
    Xu, Jin
    Peng, Deguang
    [J]. 2019 CHINESE AUTOMATION CONGRESS (CAC2019), 2019, : 3009 - 3013
  • [10] Hierarchical Gated Convolutional Networks with Multi-Head Attention for Text Classification
    Du, Haizhou
    Qian, Jingu
    [J]. 2018 5TH INTERNATIONAL CONFERENCE ON SYSTEMS AND INFORMATICS (ICSAI), 2018, : 1170 - 1175