Self-Paced Co-Training of Graph Neural Networks for Semi-Supervised Node Classification

被引:22
|
作者
Gong, Maoguo [1 ]
Zhou, Hui [1 ]
Qin, A. K. [2 ]
Liu, Wenfeng [1 ]
Zhao, Zhongying [3 ]
机构
[1] Xidian Univ, Sch Elect Engn, Xian 710071, Peoples R China
[2] Swinburne Univ Technol, Dept Comp Technol, Melbourne, Vic 3122, Australia
[3] Shandong Univ Sci & Technol, Sch Comp Sci & Engn, Qingdao 266590, Peoples R China
基金
澳大利亚研究理事会; 中国国家自然科学基金;
关键词
Training; Data models; Task analysis; Graph neural networks; Training data; Predictive models; Optimization; Co-training; graph neural networks (GNNs); node classification; self-paced learning (SPL); semi-supervised learning (SSL); COMMUNITY DETECTION;
D O I
10.1109/TNNLS.2022.3157688
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph neural networks (GNNs) have demonstrated great success in many graph data-based applications. The impressive behavior of GNNs typically relies on the availability of a sufficient amount of labeled data for model training. However, in practice, obtaining a large number of annotations is prohibitively labor-intensive and even impossible. Co-training is a popular semi-supervised learning (SSL) paradigm, which trains multiple models based on a common training set while augmenting the limited amount of labeled data used for training each model via the pseudolabeled data generated from the prediction results of other models. Most of the existing co-training works do not control the quality of pseudolabeled data when using them. Therefore, the inaccurate pseudolabels generated by immature models in the early stage of the training process are likely to cause noticeable errors when they are used for augmenting the training data for other models. To address this issue, we propose a self-paced co-training for the GNN (SPC-GNN) framework for semi-supervised node classification. This framework trains multiple GNNs with the same or different structures on different representations of the same training data. Each GNN carries out SSL by using both the originally available labeled data and the augmented pseudolabeled data generated from other GNNs. To control the quality of pseudolabels, a self-paced label augmentation strategy is designed to make the pseudolabels generated at a higher confidence level to be utilized earlier during training such that the negative impact of inaccurate pseudolabels on training data augmentation, and accordingly, the subsequent training process can be mitigated. Finally, each of the trained GNN is evaluated on a validation set, and the best-performing one is chosen as the output. To improve the training effectiveness of the framework, we devise a pretraining followed by a two-step optimization scheme to train GNNs. Experimental results on the node classification task demonstrate that the proposed framework achieves significant improvement over the state-of-the-art SSL methods.
引用
收藏
页码:9234 / 9247
页数:14
相关论文
共 50 条
  • [31] Safe co-training for semi-supervised regression
    Liu, Liyan
    Huang, Peng
    Yu, Hong
    Min, Fan
    INTELLIGENT DATA ANALYSIS, 2023, 27 (04) : 959 - 975
  • [32] Co-training with Clustering for the Semi-supervised Classification of Remote Sensing Images
    Aydav, Prem Shankar Singh
    Minz, Sonjharia
    PROCEEDINGS OF THE SECOND INTERNATIONAL CONFERENCE ON COMPUTER AND COMMUNICATION TECHNOLOGIES, IC3T 2015, VOL 2, 2016, 380 : 659 - 667
  • [33] Question classification based on co-training style semi-supervised learning
    Yu, Zhengtao
    Su, Lei
    Li, Lina
    Zhao, Quan
    Mao, Cunli
    Guo, Jianyi
    PATTERN RECOGNITION LETTERS, 2010, 31 (13) : 1975 - 1980
  • [34] Semi-Supervised and Self-Supervised Classification with Multi-View Graph Neural Networks
    Yuan, Jinliang
    Yu, Hualei
    Cao, Meng
    Xu, Ming
    Xie, Junyuan
    Wang, Chongjun
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 2466 - 2476
  • [35] Semi-Supervised Text Classification via Self-Paced Semantic-Level Contrast
    Xia, Yu
    Zhang, Kai
    Zhou, Kaijie
    Wang, Rui
    Hui, Xiaohui
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2023, PT II, 2023, 13936 : 482 - 494
  • [36] GCT: Graph Co-Training for Semi-Supervised Few-Shot Learning
    Xu, Rui
    Xing, Lei
    Shao, Shuai
    Zhao, Lifei
    Liu, Baodi
    Liu, Weifeng
    Zhou, Yicong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (12) : 8674 - 8687
  • [37] Bayesian Graph Convolutional Neural Networks for Semi-Supervised Classification
    Zhang, Yingxue
    Pal, Soumyasundar
    Coates, Mark
    Ustebay, Deniz
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 5829 - 5836
  • [38] Semi-supervised node classification via adaptive graph smoothing networks
    Zheng, Ruigang
    Chen, Weifu
    Feng, Guocan
    PATTERN RECOGNITION, 2022, 124
  • [39] SEMI-SUPERVISED CO-TRAINING AND ACTIVE LEARNING FRAMEWORK FOR HYPERSPECTRAL IMAGE CLASSIFICATION
    Samiappan, Sathishkumar
    Moorhead, Robert J., II
    2015 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS), 2015, : 401 - 404
  • [40] Development of Co-training Support Vector Machine Model for Semi-supervised Classification
    Chen, Yinghao
    Pan, Tianhong
    Chen, Shan
    PROCEEDINGS OF THE 36TH CHINESE CONTROL CONFERENCE (CCC 2017), 2017, : 11077 - 11080