Label-Guided Dual-Attention Deep Neural Network Model

被引:0
|
作者
Peng Z. [1 ]
Zhu X. [1 ]
Guo J. [2 ]
机构
[1] College of Computer Science and Engineering, Chongqing University of Technology, Chongqing
[2] Key Laboratory of Network Data Science and Technology, Institute of Computing Technology, Chinese Academy of Sciences, Beijing
基金
中国国家自然科学基金;
关键词
Dual-Attention; Label Attention; Label Description Generation; Self-Attention; Sentiment Classification;
D O I
10.16451/j.cnki.issn1003-6059.202202008
中图分类号
学科分类号
摘要
Since the text information of labels is not included in some datasets, the semantic relationship between text words and labels cannot be explicitly calculated in the existing explicit interactive classification models. To solve this problem, a label-guided dual-attention deep neural network model is proposed in this paper. Firstly, an automatic category label description generation method based on inverse label frequency is proposed. According to the label description generation method, a specific label description for each label is generated. The generated specific label description is applied to explicitly calculate the semantic relationship between text words and labels. On the basis of the above, review text representation with contextual information is learned by a text encoder. A label-guided dual-attention network is proposed to learn the text representation based on self-attention and the text representation based on label attention, respectively. Then, an adaptive gating mechanism is employed to fuse two mentioned text representations and the final text representation is thus obtained. Finally, a two-layer feedforward neural network is utilized as a classifier for sentiment classification. Experiments on three publicly available real-world datasets show that the proposed model produces better classification performance. © 2022, Science Press. All right reserved.
引用
收藏
页码:175 / 184
页数:9
相关论文
共 33 条
  • [1] HE L, FANG W L, ZHANG H Y., Aspect-Level Sentiment Classification Model Based on Context-Preserving Capability, Pattern Re-cognition and Artificial Intelligence, 34, 2, pp. 157-166, (2021)
  • [2] ZHANG L, WANG S, LIU B., Deep Learning for Sentiment Analysis: A Survey, WIREs: Data Mining and Knowledge Discovery, 8, 4, (2018)
  • [3] QIN Q, HU W P, LIU B., Feature Projection for Improved Text Classification, Proc of the 58th Annual Meeting of the Association for Computational Linguistics, pp. 8161-8171, (2020)
  • [4] WANG H, WANG S, MAZUMDER S, Et al., Bayes-Enhanced Lifelong Attention Networks for Sentiment Classification, Proc of the 28th International Conference on Computational Linguistics, pp. 580-591, (2020)
  • [5] GAO W L, YOSHINAGA N, KAJI N, Et al., Modeling User Leniency and Product Popularity for Sentiment Classification, Proc of the 6th International Joint Conference on Natural Language Processing, pp. 1107-1111, (2013)
  • [6] QU L Z, IFRIM G, WEIKUM G., The Bag-of-Opinions Method for Review Rating Prediction from Sparse Text Patterns, Proc of the 23rd International Conference on Computational Linguistics, pp. 913-921, (2010)
  • [7] GOLDBERG A B, ZHU X J., Seeing Stars When There Aren't Many Stars: Graph-Based Semi-Supervised Learning for Sentiment Categorization, Proc of the 1st Workshop on Graph Based Methods for Natural Language Processing, pp. 45-52, (2006)
  • [8] KURATA G, XIANG B, ZHOU B W., Improved Neural Network-Based Multi-label Classification with Better Initialization Leveraging Label Co-occurrence, Proc of the Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pp. 521-526, (2016)
  • [9] LAI S W, XU L H, LIU K, Et al., Recurrent Convolutional Neural Networks for Text Classification, Proc of the 29th AAAI Confe-rence on Artificial Intelligence, pp. 2267-2273, (2015)
  • [10] TANG D Y, QIN B, LIU T., Document Modeling with Gated Recurrent Neural Network for Sentiment Classification, Proc of the Conference on Empirical Methods in Natural Language Processing, pp. 1422-1432, (2015)