Understanding Dropout for Graph Neural Networks

被引:4
|
作者
Shu, Juan [1 ]
Xi, Bowei [1 ]
Li, Yu [2 ]
Wu, Fan [1 ]
Kamhoua, Charles [3 ]
Ma, Jianzhu [4 ]
机构
[1] Purdue Univ, Dept Stat, W Lafayette, IN 47907 USA
[2] Chinese Univ Hong Kong, Comp Sci & Engn, Hong Kong, Peoples R China
[3] US Army Res Lab, Adelphi, MD USA
[4] Peking Univ, Inst Artificial Intelligence, Beijing, Peoples R China
关键词
Graph neural network; dropout; over-smoothing;
D O I
10.1145/3487553.3524725
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Graph neural network (GNN) has demonstrated superior performance on graph learning tasks. GNN captures the data dependencies via message passing amid neural networks. Hence the prediction of a node label can utilize information from its neighbors in a graph. Dropout is a regularization as well as an ensemble method for convolutional neural network (CNN), which has been carefully studied. However, there are few existing works that focused on dropout schemes for GNN. Although GNN and CNN share similar model architecture, both with convolutional layers and fully connected layers, the input data structure for GNN and CNN are different and convolution operation differs. This suggests the dropout schemes for CNN should not be directly applied to GNN without a good understanding of the impact. In this paper, we divide the existing dropout schemes for GNN into two categories: (1) dropout on feature maps and (2) dropout on graph structure. Based on the drawbacks of current GNN dropout models, we propose a novel layer compensation dropout and a novel adaptive heteroscadestic Gaussian dropout, which can be applied to any type of GNN models and outperforms their corresponding baselines in shallow GNNs. Then an experimental study shows Bernoulli dropout generalize better while Gaussian dropout is slightly stronger in transductive performance. At last, we theoretically study how different dropout schemes mitigate over-smoothing problems and experimental results shows that layer compensation dropout allows a GNN model to maintain or slightly improve its performance as the GNN model adds more layers while all the other dropout models suffer from performance degradation when GNN goes deep.
引用
收藏
页码:1128 / 1138
页数:11
相关论文
共 50 条
  • [21] Graph neural networks
    不详
    [J]. NATURE REVIEWS METHODS PRIMERS, 2024, 4 (01):
  • [22] Graph Clustering with Graph Neural Networks
    Tsitsulin, Anton
    Palowitch, John
    Perozzi, Bryan
    Mueller, Emmanuel
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [23] Graph Mining with Graph Neural Networks
    Jin, Wei
    [J]. WSDM '21: PROCEEDINGS OF THE 14TH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2021, : 1119 - 1120
  • [24] Graph Neural Networks for Graph Drawing
    Tiezzi, Matteo
    Ciravegna, Gabriele
    Gori, Marco
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (04) : 4668 - 4681
  • [25] Towards dropout training for convolutional neural networks
    Wu, Haibing
    Gu, Xiaodong
    [J]. NEURAL NETWORKS, 2015, 71 : 1 - 10
  • [26] Graphs, Convolutions, and Neural Networks: From Graph Filters to Graph Neural Networks
    Gama, Fernando
    Isufi, Elvin
    Leus, Geert
    Ribeiro, Alejandro
    [J]. IEEE SIGNAL PROCESSING MAGAZINE, 2020, 37 (06) : 128 - 138
  • [27] Variational Dropout Sparsifies Deep Neural Networks
    Molchanov, Dmitry
    Ashukha, Arsenii
    Vetrov, Dmitry
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [28] Dropout Rademacher complexity of deep neural networks
    Wei GAO
    Zhi-Hua ZHOU
    [J]. Science China(Information Sciences), 2016, 59 (07) : 173 - 184
  • [29] Regularization of deep neural networks with spectral dropout
    Khan, Salman H.
    Hayat, Munawar
    Porikli, Fatih
    [J]. NEURAL NETWORKS, 2019, 110 : 82 - 90
  • [30] Augmenting Recurrent Neural Networks Resilience by Dropout
    Bacciu, Davide
    Crecchi, Francesco
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (01) : 345 - 351