Graph-based Text Classification by Contrastive Learning with Text-level Graph Augmentation

被引:3
|
作者
Li, Ximing [1 ,2 ]
Wang, Bing [1 ,2 ]
Wang, Yang [3 ]
Wang, Meng [3 ]
机构
[1] Jilin Univ, Coll Comp Sci & Technol, Jilin, Jilin, Peoples R China
[2] Jilin Univ, Minist Educ, Key Lab Symbol Comp & Knowledge Engn, Jilin, Jilin, Peoples R China
[3] Hefei Univ Technol, Sch Comp Sci & Informat Engn, Hefei, Peoples R China
基金
中国国家自然科学基金;
关键词
Multi-label classification; graph representation; label correlation; contrastive learning; graph augmentation;
D O I
10.1145/3638353
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Text Classification (TC) is a fundamental task in the information retrieval community. Nowadays, the mainstay TC methods are built on the deep neural networks, which can learn much more discriminative text features than the traditional shallow learning methods. Among existing deep TC methods, the ones based on Graph Neural Network (GNN) have attracted more attention due to the superior performance. Technically, the GNN-based TC methods mainly transform the full training dataset to a graph of texts; however, they often neglect the dependency between words, so as to miss potential semantic information of texts, which may be significant to exactly represent them. To solve the aforementioned problem, we generate graphs of words instead, so as to capture the dependency information of words. Specifically, each text is translated into a graph of words, where neighboring words are linked. We learn the node features of words by a GNN-like procedure and then aggregate them as the graph feature to represent the current text. To further improve the text representations, we suggest a contrastive learning regularization term. Specifically, we generate two augmented text graphs for each original text graph, we constrain the representations of the two augmented graphs from the same text close and the ones from different texts far away. We propose various techniques to generate the augmented graphs. Upon those ideas, we develop a novel deep TC model, namely Text-level Graph Networks with Contrastive Learning (TGNcl). We conduct a number of experiments to evaluate the proposed TGNcl model. The empirical results demonstrate that TGNcl can outperform the existing state-of-the-art TC models.
引用
收藏
页数:21
相关论文
共 50 条
  • [1] Text-Level Contrastive Learning for Scene Text Recognition
    Zhuang, Junbin
    Ren, Yixuan
    Li, Xia
    Liang, Zhanpeng
    [J]. 2022 INTERNATIONAL CONFERENCE ON ASIAN LANGUAGE PROCESSING (IALP 2022), 2022, : 231 - 236
  • [2] Contrastive Graph Convolutional Networks with adaptive augmentation for text classification
    Yang, Yintao
    Miao, Rui
    Wang, Yili
    Wang, Xin
    [J]. INFORMATION PROCESSING & MANAGEMENT, 2022, 59 (04)
  • [3] Contrastive learning with text augmentation for text classification
    Jia, Ouyang
    Huang, Huimin
    Ren, Jiaxin
    Xie, Luodi
    Xiao, Yinyin
    [J]. APPLIED INTELLIGENCE, 2023, 53 (16) : 19522 - 19531
  • [4] Contrastive learning with text augmentation for text classification
    Ouyang Jia
    Huimin Huang
    Jiaxin Ren
    Luodi Xie
    Yinyin Xiao
    [J]. Applied Intelligence, 2023, 53 : 19522 - 19531
  • [5] TextGCL: Graph Contrastive Learning for Transductive Text Classification
    Zhao, Yawei
    Song, Xiaoyang
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [6] Improved Graph Contrastive Learning for Short Text Classification
    Liu, Yonghao
    Huang, Lan
    Giunchiglia, Fausto
    Feng, Xiaoyue
    Guan, Renchu
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 18716 - 18724
  • [7] Graph-based Semi-supervised Learning for Text Classification
    Widmann, Natalie
    Verberne, Suzan
    [J]. ICTIR'17: PROCEEDINGS OF THE 2017 ACM SIGIR INTERNATIONAL CONFERENCE THEORY OF INFORMATION RETRIEVAL, 2017, : 59 - 66
  • [8] A Graph-Based Measurement for Text Imbalance Classification
    Tian, Jiachen
    Chen, Shizhan
    Zhang, Xiaowang
    Feng, Zhiyong
    [J]. ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 2188 - 2195
  • [9] KGAT: An Enhanced Graph-Based Model for Text Classification
    Wang, Xin
    Wang, Chao
    Yang, Haiyang
    Zhang, Xingpeng
    Shen, Qi
    Ji, Kan
    Wu, Yuhong
    Zhan, Huayi
    [J]. NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2022, PT I, 2022, 13551 : 656 - 668
  • [10] Contrastive Learning with Heterogeneous Graph Attention Networks on Short Text Classification
    Sun, Zhongtian
    Harit, Anoushka
    Cristea, Alexandra, I
    Yu, Jialin
    Shi, Lei
    Al Moubayed, Noura
    [J]. 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,