Self-Paced Co-Training of Graph Neural Networks for Semi-Supervised Node Classification

被引:22
|
作者
Gong, Maoguo [1 ]
Zhou, Hui [1 ]
Qin, A. K. [2 ]
Liu, Wenfeng [1 ]
Zhao, Zhongying [3 ]
机构
[1] Xidian Univ, Sch Elect Engn, Xian 710071, Peoples R China
[2] Swinburne Univ Technol, Dept Comp Technol, Melbourne, Vic 3122, Australia
[3] Shandong Univ Sci & Technol, Sch Comp Sci & Engn, Qingdao 266590, Peoples R China
基金
澳大利亚研究理事会; 中国国家自然科学基金;
关键词
Training; Data models; Task analysis; Graph neural networks; Training data; Predictive models; Optimization; Co-training; graph neural networks (GNNs); node classification; self-paced learning (SPL); semi-supervised learning (SSL); COMMUNITY DETECTION;
D O I
10.1109/TNNLS.2022.3157688
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph neural networks (GNNs) have demonstrated great success in many graph data-based applications. The impressive behavior of GNNs typically relies on the availability of a sufficient amount of labeled data for model training. However, in practice, obtaining a large number of annotations is prohibitively labor-intensive and even impossible. Co-training is a popular semi-supervised learning (SSL) paradigm, which trains multiple models based on a common training set while augmenting the limited amount of labeled data used for training each model via the pseudolabeled data generated from the prediction results of other models. Most of the existing co-training works do not control the quality of pseudolabeled data when using them. Therefore, the inaccurate pseudolabels generated by immature models in the early stage of the training process are likely to cause noticeable errors when they are used for augmenting the training data for other models. To address this issue, we propose a self-paced co-training for the GNN (SPC-GNN) framework for semi-supervised node classification. This framework trains multiple GNNs with the same or different structures on different representations of the same training data. Each GNN carries out SSL by using both the originally available labeled data and the augmented pseudolabeled data generated from other GNNs. To control the quality of pseudolabels, a self-paced label augmentation strategy is designed to make the pseudolabels generated at a higher confidence level to be utilized earlier during training such that the negative impact of inaccurate pseudolabels on training data augmentation, and accordingly, the subsequent training process can be mitigated. Finally, each of the trained GNN is evaluated on a validation set, and the best-performing one is chosen as the output. To improve the training effectiveness of the framework, we devise a pretraining followed by a two-step optimization scheme to train GNNs. Experimental results on the node classification task demonstrate that the proposed framework achieves significant improvement over the state-of-the-art SSL methods.
引用
收藏
页码:9234 / 9247
页数:14
相关论文
共 50 条
  • [1] Self-paced and self-consistent co-training for semi-supervised image segmentation
    Wang, Ping
    Peng, Jizong
    Pedersoli, Marco
    Zhou, Yuanfeng
    Zhang, Caiming
    Desrosiers, Christian
    MEDICAL IMAGE ANALYSIS, 2021, 73
  • [2] Adaptive Graph Learning for Semi-supervised Self-paced Classification
    Long Chen
    Jianbo Lu
    Neural Processing Letters, 2022, 54 : 2695 - 2716
  • [3] Adaptive Graph Learning for Semi-supervised Self-paced Classification
    Chen, Long
    Lu, Jianbo
    NEURAL PROCESSING LETTERS, 2022, 54 (04) : 2695 - 2716
  • [4] Classwise Self-Paced Self-Training for Semi-Supervised Image Classification
    Lu, Cheng-Yu
    Hsu, Heng-Cheng
    Chiang, Chen-Kuo
    2023 ASIA PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE, APSIPA ASC, 2023, : 753 - 758
  • [5] Self-Paced Co-training
    Ma, Fan
    Meng, Deyu
    Xie, Qi
    Li, Zina
    Dong, Xuanyi
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [6] Self-Consistent Graph Neural Networks for Semi-Supervised Node Classification
    Liu, Yanbei
    Zhao, Shichuan
    Wang, Xiao
    Geng, Lei
    Xiao, Zhitao
    Lin, Jerry Chun-Wei
    IEEE TRANSACTIONS ON BIG DATA, 2023, 9 (04) : 1186 - 1197
  • [7] Semi-Supervised Image Classification With Self-Paced Cross-Task Networks
    Wu, Si
    Ji, Qiujia
    Wang, Shufeng
    Wong, Hau-San
    Yu, Zhiwen
    Xu, Yong
    IEEE TRANSACTIONS ON MULTIMEDIA, 2018, 20 (04) : 851 - 865
  • [8] Enhancing Graph Neural Networks via auxiliary training for semi-supervised node classification
    Wu, Yao
    Song, Yu
    Huang, Hong
    Ye, Fanghua
    Xie, Xing
    Jin, Hai
    KNOWLEDGE-BASED SYSTEMS, 2021, 220
  • [9] Co-training generative adversarial networks for semi-supervised classification method
    Xu, Zhe
    Geng, Jie
    Jiang, Wen
    Zhang, Zhuo
    Zeng, Qing-Jie
    Guangxue Jingmi Gongcheng/Optics and Precision Engineering, 2021, 29 (05): : 1127 - 1135
  • [10] Semi-Supervised Classification with Co-training for Deep Web
    Fang Wei
    Cui Zhiming
    ADVANCED MEASUREMENT AND TEST, PARTS 1 AND 2, 2010, 439-440 : 183 - +