Dual-axial self-attention network for text classification

被引:7
|
作者
Zhang, Xiaochuan [1 ]
Qiu, Xipeng [2 ,3 ]
Pang, Jianmin [1 ]
Liu, Fudong [1 ]
Li, Xingwei [1 ]
机构
[1] State Key Lab Math Engn & Adv Comp, Zhengzhou 450001, Peoples R China
[2] Fudan Univ, Shanghai Key Lab Intelligent Informat Proc, Shanghai 201203, Peoples R China
[3] Fudan Univ, Sch Comp Sci, Shanghai 201203, Peoples R China
基金
中国国家自然科学基金;
关键词
text classification; dual-axial self-attention; feature-axial dependency;
D O I
10.1007/s11432-019-2744-2
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Text classification is an important task in natural language processing and numerous studies aim to improve the accuracy and efficiency of text classification models. In this study, we propose an effective and efficient text classification model which is based on self-attention solely. The recently proposed multi-dimensional self-attention significantly improved the performance of self-attention. However, existing models suffer from two major limitations: (1) the previous multi-dimensional self-attention models are quite time-consuming; (2) the dependencies of elements along the feature axis are not taken into account. To overcome these problems, in this paper, a much more computational efficient multi-dimensional self-attention model is proposed, and two parallel self-attention modules, called dual-axial self-attention, are applied to capture rich dependencies along the feature axis as well as the text axis. A text classification model is then derived. The experimental results on eight representative datasets show that the proposed text classification model can obtain state-of-the-art results and the proposed self-attention outperforms conventional self-attention models.
引用
下载
收藏
页数:11
相关论文
共 50 条
  • [1] Dual-axial self-attention network for text classification
    Xiaochuan Zhang
    Xipeng Qiu
    Jianmin Pang
    Fudong Liu
    Xingwei Li
    Science China Information Sciences, 2021, 64
  • [2] Dual-axial self-attention network for text classification
    Xiaochuan ZHANG
    Xipeng QIU
    Jianmin PANG
    Fudong LIU
    Xingwei LI
    Science China(Information Sciences), 2021, 64 (12) : 80 - 90
  • [3] A Self-attention Based LSTM Network for Text Classification
    Jing, Ran
    2019 3RD INTERNATIONAL CONFERENCE ON CONTROL ENGINEERING AND ARTIFICIAL INTELLIGENCE (CCEAI 2019), 2019, 1207
  • [4] Multiple Positional Self-Attention Network for Text Classification
    Dai, Biyun
    Li, Jinlong
    Xu, Ruoyi
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 7610 - 7617
  • [5] Deformable Self-Attention for Text Classification
    Ma, Qianli
    Yan, Jiangyue
    Lin, Zhenxi
    Yu, Liuhong
    Chen, Zipeng
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2021, 29 : 1570 - 1581
  • [6] DCCL: Dual-channel hybrid neural network combined with self-attention for text classification
    Li, Chaofan
    Qiong, Liu
    Kai, Ma
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2023, 20 (02) : 1981 - 1992
  • [7] Research on a Capsule Network Text Classification Method with a Self-Attention Mechanism
    Yu, Xiaodong
    Luo, Shun-Nain
    Wu, Yujia
    Cai, Zhufei
    Kuan, Ta-Wen
    Tseng, Shih-Pang
    SYMMETRY-BASEL, 2024, 16 (05):
  • [8] Multi-Scale Self-Attention for Text Classification
    Guo, Qipeng
    Qiu, Xipeng
    Liu, Pengfei
    Xue, Xiangyang
    Zhang, Zheng
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 7847 - 7854
  • [9] Quantum self-attention neural networks for text classification
    Li, Guangxi
    Zhao, Xuanqiang
    Wang, Xin
    SCIENCE CHINA-INFORMATION SCIENCES, 2024, 67 (04)
  • [10] Quantum self-attention neural networks for text classification
    Guangxi LI
    Xuanqiang ZHAO
    Xin WANG
    Science China(Information Sciences), 2024, 67 (04) : 301 - 313