Generalizing Graph Neural Networks on Out-of-Distribution Graphs

被引:5
|
作者
Fan, Shaohua [1 ,2 ]
Wang, Xiao [3 ]
Shi, Chuan [1 ]
Cui, Peng [2 ]
Wang, Bai [1 ]
机构
[1] Beijing Univ Posts & Telecommun, Dept Comp Sci, Beijing 100876, Peoples R China
[2] Tsinghua Univ, Dept Comp Sci & Technol, Beijing 100084, Peoples R China
[3] Beihang Univ, Sch Software, Beijing 100191, Peoples R China
基金
中国国家自然科学基金;
关键词
Causal representation learning; graph neural networks (GNNs); out-of-distribution generalization (OOD); stable learning; DEPENDENCE;
D O I
10.1109/TPAMI.2023.3321097
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph Neural Networks (GNNs) are proposed without considering the agnostic distribution shifts between training graphs and testing graphs, inducing the degeneration of the generalization ability of GNNs in Out-Of-Distribution (OOD) settings. The fundamental reason for such degeneration is that most GNNs are developed based on the I.I.D hypothesis. In such a setting, GNNs tend to exploit subtle statistical correlations existing in the training set for predictions, even though it is a spurious correlation. This learning mechanism inherits from the common characteristics of machine learning approaches. However, such spurious correlations may change in the wild testing environments, leading to the failure of GNNs. Therefore, eliminating the impact of spurious correlations is crucial for stable GNN models. To this end, in this paper, we argue that the spurious correlation exists among subgraph-level units and analyze the degeneration of GNN in causal view. Based on the causal view analysis, we propose a general causal representation framework for stable GNN, called StableGNN. The main idea of this framework is to extract high-level representations from raw graph data first and resort to the distinguishing ability of causal inference to help the model get rid of spurious correlations. Particularly, to extract meaningful high-level representations, we exploit a differentiable graph pooling layer to extract subgraph-based representations by an end-to-end manner. Furthermore, inspired by the confounder balancing techniques from causal inference, based on the learned high-level representations, we propose a causal variable distinguishing regularizer to correct the biased training distribution by learning a set of sample weights. Hence, GNNs would concentrate more on the true connection between discriminative substructures and labels. Extensive experiments are conducted on both synthetic datasets with various distribution shift degrees and eight real-world OOD graph datasets. The results well verify that the proposed model StableGNN not only outperforms the state-of-the-arts but also provides a flexible framework to enhance existing GNNs. In addition, the interpretability experiments validate that StableGNN could leverage causal structures for predictions.
引用
收藏
页码:322 / 337
页数:16
相关论文
共 50 条
  • [1] MaxEnt Loss: Calibrating Graph Neural Networks under Out-of-Distribution Shift (Student Abstract)
    Neo, Dexter
    [J]. THIRTY-EIGTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 21, 2024, : 23594 - 23596
  • [2] A Data-centric Framework to Endow Graph Neural Networks with Out-Of-Distribution Detection Ability
    Guo, Yuxin
    Yang, Cheng
    Chen, Yuluo
    Liu, Jixi
    Shi, Chuan
    Du, Junping
    [J]. PROCEEDINGS OF THE 29TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2023, 2023, : 638 - 648
  • [3] Learning on Graphs with Out-of-Distribution Nodes
    Song, Yu
    Wang, Donglin
    [J]. PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 1635 - 1645
  • [4] OOD-GNN: Out-of-Distribution Generalized Graph Neural Network
    Li, Haoyang
    Wang, Xin
    Zhang, Ziwei
    Zhu, Wenwu
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (07) : 7328 - 7340
  • [5] Layer Adaptive Deep Neural Networks for Out-of-Distribution Detection
    Wang, Haoliang
    Zhao, Chen
    Zhao, Xujiang
    Chen, Feng
    [J]. ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2022, PT II, 2022, 13281 : 526 - 538
  • [6] NeuralFP: Out-of-distribution Detection using Fingerprints of Neural Networks
    Lee, Wei-Han
    Millman, Steve
    Desai, Nirmit
    Srivatsa, Mudhakar
    Liu, Changchang
    [J]. 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 9561 - 9568
  • [7] Performance analysis of out-of-distribution detection on trained neural networks
    Henriksson, Jens
    Berger, Christian
    Borg, Markus
    Tornberg, Lars
    Sathyamoorthy, Sankar Raman
    Englund, Cristofer
    [J]. INFORMATION AND SOFTWARE TECHNOLOGY, 2021, 130
  • [8] Fixing Robust Out-of-distribution Detection for Deep Neural Networks
    Zhou, Zhiyang
    Liu, Jie
    Dou, Wensheng
    Li, Shuo
    Kang, Liangyi
    Qu, Muzi
    Ye, Dan
    [J]. 2023 IEEE 34TH INTERNATIONAL SYMPOSIUM ON SOFTWARE RELIABILITY ENGINEERING, ISSRE, 2023, : 533 - 544
  • [9] Calibrate Graph Neural Networks under Out-of-Distribution Nodes via Deep Q-learning
    Shi, Weili
    Yang, Xueying
    Zhao, Xujiang
    Chen, Haifeng
    Tao, Zhiqiang
    Li, Sheng
    [J]. PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 2270 - 2279
  • [10] Generalizing Graph Convolutional Neural Networks with Edge-Variant Recursions on Graphs
    Isufi, Elvin
    Gama, Fernando
    Ribeiro, Alejandro
    [J]. 2019 27TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2019,