A two-stage co-adversarial perturbation to mitigate out-of-distribution generalization of large-scale graph

被引:0
|
作者
Wang, Yili [1 ]
Xue, Haotian [1 ]
Wang, Xin [1 ]
机构
[1] Jilin Univ, Sch Artificial Intelligence, Changchun 130012, Peoples R China
基金
中国国家自然科学基金;
关键词
Graph neural network; Adversarial training; Graph out-of-distribution; NETWORK;
D O I
10.1016/j.eswa.2024.124472
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In the realm of graph out-of-distribution (OOD), despite recent strides in advancing graph neural networks (GNNs) for the modeling of graph data, training GNNs on large-scale datasets presents a formidable hurdle due to the pervasive challenge of overfitting. To address these issues, researchers have explored adversarial training, a technique that enriches training data with worst-case adversarial examples. However, while prior work on adversarial training primarily focuses on safeguarding GNNs against malicious attacks, its potential to enhance the OOD generalization abilities of GNNs in the context of graph analytics remains less explored. In our research, we delve into the inner workings of GNNs by examining the landscapes of weight and feature losses, which respectively illustrate how the loss function changes concerning model weights and node features. Our investigation reveals a noteworthy phenomenon: GNNs are inclined to become trapped in sharp local minima within these loss landscapes, resulting in suboptimal OOD generalization performance. To address this challenge, we introduce the concept of co-adversarial perturbation optimization, which considers both model weights and node features, and we design an alternating adversarial perturbation algorithm for graph out-of-distribution generalization. This algorithm operates iteratively, smoothing the weight and feature loss landscapes alternately. Moreover, our training process unfolds in two distinct stages. The first stage centers on standard cross-entropy minimization, ensuring rapid convergence of GNN models. In the second stage, we employ our alternating adversarial training strategy to prevent the models from becoming ensnared in locally sharp minima. Our extensive experiments provide compelling evidence that our CAP approach can generally enhance the OOD generalization performance of GNNs across a diverse range of large-scale graphs.
引用
收藏
页数:11
相关论文
共 50 条
  • [11] Decomposition algorithm for large-scale two-stage unit-commitment
    van Ackooij, Wim
    Malick, Jerome
    ANNALS OF OPERATIONS RESEARCH, 2016, 238 (1-2) : 587 - 613
  • [12] On using stratified two-stage sampling for large-scale multispecies surveys
    Aubry, Philippe
    Quaintenne, Gwenael
    Dupuy, Jeremy
    Francesiaz, Charlotte
    Guillemain, Matthieu
    Caizergues, Alain
    ECOLOGICAL INFORMATICS, 2023, 77
  • [13] A two-stage optimization strategy for large-scale oil field development
    Nasir, Yusuf
    Volkov, Oleg
    Durlofsky, Louis J.
    OPTIMIZATION AND ENGINEERING, 2022, 23 (01) : 361 - 395
  • [14] Two-Stage Optimal Scheduling Strategy for Large-Scale Electric Vehicles
    Wang X.
    Sun C.
    Wang R.
    Wei T.
    Wang, Rutian (wrtmail@163.com), 1600, Institute of Electrical and Electronics Engineers Inc., United States (08): : 13821 - 13832
  • [16] Two-stage based Ensemble Optimization for Large-Scale Global Optimization
    Wang, Yu
    Li, Bin
    2010 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2010,
  • [17] A Two-Stage Fault Tolerance Method for Large-Scale Manufacturing Network
    Wu, Yinan
    Peng, Gongzhuang
    Wang, Hongwei
    Zhang, Heming
    IEEE ACCESS, 2019, 7 : 81574 - 81592
  • [18] Decomposition algorithm for large-scale two-stage unit-commitment
    Wim van Ackooij
    Jérôme Malick
    Annals of Operations Research, 2016, 238 : 587 - 613
  • [19] A two-stage optimization strategy for large-scale oil field development
    Yusuf Nasir
    Oleg Volkov
    Louis J. Durlofsky
    Optimization and Engineering, 2022, 23 : 361 - 395
  • [20] Two-Stage Nonnegative Sparse Representation for Large-Scale Face Recognition
    He, Ran
    Zheng, Wei-Shi
    Hu, Bao-Gang
    Kong, Xiang-Wei
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2013, 24 (01) : 35 - 46