Multi-label sequence generating model via label semantic attention mechanism

被引:2
|
作者
Zhang, Xiuling [1 ]
Tan, Xiaofei [1 ]
Luo, Zhaoci [1 ]
Zhao, Jun [1 ]
机构
[1] Yanshan Univ, Engn Res Ctr, Minist Educ Intelligent Control Syst & Intelligen, Qinhuangdao 066000, Hebei, Peoples R China
关键词
Multi-label text classification; Seq2Seq; Label semantic attention mechanism; Policy gradient; NEURAL-NETWORKS; TEXT;
D O I
10.1007/s13042-022-01722-4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, a new attempt has been made to capture label co-occurrence by applying the sequence-to-sequence (Seq2Seq) model to multi-label text classification (MLTC). However, existing approaches frequently ignore the semantic information contained in the labels themselves. Besides, the Seq2Seq model is susceptible to the negative impact of label sequence order. Furthermore, it has been demonstrated that the traditional attention mechanism underperforms in MLTC. Therefore, we propose a novel Seq2Seq model with a different label semantic attention mechanism (S2S-LSAM), which generates fused information containing label and text information through the interaction of label semantics and text features in the label semantic attention mechanism. With the fused information, our model can select the text features that are most relevant to the labels more effectively. A combination of the cross-entropy loss function and the policy gradient-based loss function is employed to reduce the label sequence order effect. The experiments show that our model outperforms the baseline models.
引用
收藏
页码:1711 / 1723
页数:13
相关论文
共 50 条
  • [41] A multi-scale semantic attention representation for multi-label image recognition with graph networks
    Liang, Jun
    Xu, Feiteng
    Yu, Songsen
    Neurocomputing, 2022, 491 : 14 - 23
  • [42] Multi-label Learning via Codewords
    Sedghi, Mahlagha
    Huang, Yinjie
    Georgiopoulos, Michael
    Anagnostopoulos, Georgios
    2018 IEEE 30TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI), 2018, : 221 - 228
  • [43] Semantic-Aware Graph Matching Mechanism for Multi-Label Image Recognition
    Wu, Yanan
    Feng, Songhe
    Wang, Yang
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (11) : 6788 - 6803
  • [44] Multi-label Text Classification Model Combining BiLSTM and Hypergraph Attention
    Wang, Xing
    Hu, HuiTing
    Zhu, GuoHua
    2024 4TH INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATION AND ARTIFICIAL INTELLIGENCE, CCAI 2024, 2024, : 344 - 349
  • [45] From Softmax to Sparsemax: A Sparse Model of Attention and Multi-Label Classification
    Martins, Andre F. T.
    Astudillo, Ramon F.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [46] Towards Unbiased Multi-Label Zero-Shot Learning With Pyramid and Semantic Attention
    Liu, Ziming
    Guo, Song
    Guo, Jingcai
    Xu, Yuanyuan
    Huo, Fushuo
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 7441 - 7455
  • [47] Improved sequence generation model for multi-label classification via CNN and initialized fully connection
    Liao, Weizhi
    Wang, Yu
    Yin, Yanchao
    Zhang, Xiaobing
    Ma, Pan
    NEUROCOMPUTING, 2020, 382 (188-195) : 188 - 195
  • [48] Label-text bi-attention capsule networks model for multi-label text classification
    Wang, Gang
    Du, Yajun
    Jiang, Yurui
    Liu, Jia
    Li, Xianyong
    Chen, Xiaoliang
    Gao, Hongmei
    Xie, Chunzhi
    Lee, Yan-li
    NEUROCOMPUTING, 2024, 588
  • [49] Generating Multi-label Adversarial Examples by Linear Programming
    Zhou, Nan
    Luo, Wenjian
    Lin, Xin
    Xu, Peilan
    Zhang, Zhenya
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [50] Research of multi-label text classification based on label attention and correlation networks
    Yuan, Ling
    Xu, Xinyi
    Sun, Ping
    Yu, Hai ping
    Wei, Yin Zhen
    Zhou, Jun jie
    PLOS ONE, 2024, 19 (09):