Self-Paced Co-Training of Graph Neural Networks for Semi-Supervised Node Classification

被引:22
|
作者
Gong, Maoguo [1 ]
Zhou, Hui [1 ]
Qin, A. K. [2 ]
Liu, Wenfeng [1 ]
Zhao, Zhongying [3 ]
机构
[1] Xidian Univ, Sch Elect Engn, Xian 710071, Peoples R China
[2] Swinburne Univ Technol, Dept Comp Technol, Melbourne, Vic 3122, Australia
[3] Shandong Univ Sci & Technol, Sch Comp Sci & Engn, Qingdao 266590, Peoples R China
基金
澳大利亚研究理事会; 中国国家自然科学基金;
关键词
Training; Data models; Task analysis; Graph neural networks; Training data; Predictive models; Optimization; Co-training; graph neural networks (GNNs); node classification; self-paced learning (SPL); semi-supervised learning (SSL); COMMUNITY DETECTION;
D O I
10.1109/TNNLS.2022.3157688
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph neural networks (GNNs) have demonstrated great success in many graph data-based applications. The impressive behavior of GNNs typically relies on the availability of a sufficient amount of labeled data for model training. However, in practice, obtaining a large number of annotations is prohibitively labor-intensive and even impossible. Co-training is a popular semi-supervised learning (SSL) paradigm, which trains multiple models based on a common training set while augmenting the limited amount of labeled data used for training each model via the pseudolabeled data generated from the prediction results of other models. Most of the existing co-training works do not control the quality of pseudolabeled data when using them. Therefore, the inaccurate pseudolabels generated by immature models in the early stage of the training process are likely to cause noticeable errors when they are used for augmenting the training data for other models. To address this issue, we propose a self-paced co-training for the GNN (SPC-GNN) framework for semi-supervised node classification. This framework trains multiple GNNs with the same or different structures on different representations of the same training data. Each GNN carries out SSL by using both the originally available labeled data and the augmented pseudolabeled data generated from other GNNs. To control the quality of pseudolabels, a self-paced label augmentation strategy is designed to make the pseudolabels generated at a higher confidence level to be utilized earlier during training such that the negative impact of inaccurate pseudolabels on training data augmentation, and accordingly, the subsequent training process can be mitigated. Finally, each of the trained GNN is evaluated on a validation set, and the best-performing one is chosen as the output. To improve the training effectiveness of the framework, we devise a pretraining followed by a two-step optimization scheme to train GNNs. Experimental results on the node classification task demonstrate that the proposed framework achieves significant improvement over the state-of-the-art SSL methods.
引用
收藏
页码:9234 / 9247
页数:14
相关论文
共 50 条
  • [21] Structure regularized self-paced learning for robust semi-supervised pattern classification
    Gu, Nannan
    Fan, Pengying
    Fan, Mingyu
    Wang, Di
    NEURAL COMPUTING & APPLICATIONS, 2019, 31 (10): : 6559 - 6574
  • [22] Robust Semi-Supervised Classification for Noisy Labels Based on Self-Paced Learning
    Gu, Nannan
    Fan, Mingyu
    Meng, Deyu
    IEEE SIGNAL PROCESSING LETTERS, 2016, 23 (12) : 1806 - 1810
  • [23] Self-paced multi-label co-training
    Gong, Yanlu
    Wu, Quanwang
    Zhou, Mengchu
    Wen, Junhao
    INFORMATION SCIENCES, 2023, 622 : 269 - 281
  • [24] Self-paced Multi-view Co-training
    Ma, Fan
    Meng, Deyu
    Dong, Xuanyi
    Yang, Yi
    JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [25] Discriminative Graph Convolutional Networks for Semi-supervised Node Classification
    Ai, Guoguo
    Yan, Hui
    Chen, Yuxin
    2022 IEEE 34TH INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, ICTAI, 2022, : 372 - 376
  • [26] Hierarchical graph attention networks for semi-supervised node classification
    Kangjie Li
    Yixiong Feng
    Yicong Gao
    Jian Qiu
    Applied Intelligence, 2020, 50 : 3441 - 3451
  • [27] Progressive Graph Convolutional Networks for Semi-Supervised Node Classification
    Heidari, Negar
    Iosifidis, Alexandros
    IEEE ACCESS, 2021, 9 : 81957 - 81968
  • [28] Hierarchical graph attention networks for semi-supervised node classification
    Feng, Yixiong
    Li, Kangjie
    Gao, Yicong
    Qiu, Jian
    APPLIED INTELLIGENCE, 2020, 50 (10) : 3441 - 3451
  • [29] Co-Training Based Semi-supervised Classification of Alzheimer's Disease
    Zhu, Jie
    Shi, Jun
    Liu, Xiao
    Chen, Xin
    2014 19TH INTERNATIONAL CONFERENCE ON DIGITAL SIGNAL PROCESSING (DSP), 2014, : 729 - 732
  • [30] Label-Consistency based Graph Neural Networks for Semi-supervised Node Classification
    Xu, Bingbing
    Huang, Junjie
    Hou, Liang
    Shen, Huawei
    Gao, Jinhua
    Cheng, Xueqi
    PROCEEDINGS OF THE 43RD INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '20), 2020, : 1897 - 1900