Hierarchical multi-attention networks for document classification

被引:1
|
作者
Yingren Huang
Jiaojiao Chen
Shaomin Zheng
Yun Xue
Xiaohui Hu
机构
[1] Guangdong University of Foreign Studies,Laboratory of Language Engineering and Computing
[2] South China Normal University,Guangdong Provincial Key Laboratory of Quantum Engineering and Quantum Materials, School of Physics and Telecommunication Engineering
关键词
Document classification; Hierarchical network; Bi-GRU; Attention mechanism;
D O I
暂无
中图分类号
学科分类号
摘要
Research of document classification is ongoing to employ the attention based-deep learning algorithms and achieves impressive results. Owing to the complexity of the document, classical models, as well as single attention mechanism, fail to meet the demand of high-accuracy classification. This paper proposes a method that classifies the document via the hierarchical multi-attention networks, which describes the document from the word-sentence level and the sentence-document level. Further, different attention strategies are performed on different levels, which enables accurate assigning of the attention weight. Specifically, the soft attention mechanism is applied to the word-sentence level while the CNN-attention to the sentence-document level. Due to the distinctiveness of the model, the proposed method delivers the highest accuracy compared to other state-of-the-art methods. In addition, the attention weight visualization outcomes present the effectiveness of attention mechanism in distinguishing the importance.
引用
收藏
页码:1639 / 1647
页数:8
相关论文
共 50 条
  • [11] A Classification Model of Legal Consulting Questions Based on Multi-Attention Prototypical Networks
    Feng, Jianzhou
    Cui, Jinman
    Wei, Qikai
    Zhou, Zhengji
    Wang, Yuxiong
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE SYSTEMS, 2022, 14 (01)
  • [12] A Classification Model of Legal Consulting Questions Based on Multi-Attention Prototypical Networks
    Jianzhou Feng
    Jinman Cui
    Qikai Wei
    Zhengji Zhou
    Yuxiong Wang
    International Journal of Computational Intelligence Systems, 14
  • [13] Multi-attention Fusion for Multimodal Sentiment Classification
    Li, Guangmin
    Zeng, Xin
    Chen, Chi
    Zhou, Long
    PROCEEDINGS OF 2024 ACM ICMR WORKSHOP ON MULTIMODAL VIDEO RETRIEVAL, ICMR-MVR 2024, 2024, : 1 - 7
  • [14] Parallel Hierarchical Attention Networks with Shared Memory Reader for Multi-Stream Conversational Document Classification
    Sawada, Naoki
    Masumura, Ryo
    Nishizaki, Hiromitsu
    18TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2017), VOLS 1-6: SITUATED INTERACTION, 2017, : 3311 - 3315
  • [15] Video Captioning using Hierarchical Multi-Attention Model
    Xiao, Huanhou
    Shi, Jinglun
    ICAIP 2018: 2018 THE 2ND INTERNATIONAL CONFERENCE ON ADVANCES IN IMAGE PROCESSING, 2018, : 96 - 101
  • [16] Hyperspectral Image Classification Based on Multi-attention Mechanism and Compiled Graph Neural Networks
    Jie S.
    Jing Y.
    Shujie D.
    Shaobo L.
    Jianjun H.
    Nongye Jixie Xuebao/Transactions of the Chinese Society for Agricultural Machinery, 2024, 55 (03): : 183 - 192and212
  • [17] Short Text Classification Model Based on Multi-Attention
    Liu, Yunxiang
    Xu, Qi
    2020 13TH INTERNATIONAL SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE AND DESIGN (ISCID 2020), 2020, : 225 - 229
  • [18] MULTI-ATTENTION NETWORK FOR THORACIC DISEASE CLASSIFICATION AND LOCALIZATION
    Ma, Yanbo
    Zhou, Qiuhao
    Chen, Xuesong
    Lu, Haihua
    Zhao, Yong
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 1378 - 1382
  • [19] Multi-scale Multi-attention Network for Moire Document Image Binarization
    Guo, Yanqing
    Ji, Caijuan
    Zheng, Xin
    Wang, Qianyu
    Luo, Xiangyang
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2021, 90
  • [20] Applying Siamese Hierarchical Attention Neural Networks for multi-document summarization
    Angel Gonzalez, Jose
    Delonca, Julien
    Sanchis, Emilio
    Garcia-Granada, Fernando
    Segarra, Encarna
    PROCESAMIENTO DEL LENGUAJE NATURAL, 2019, (63): : 111 - 118