Learning Graph Neural Networks with Approximate Gradient Descent

被引:0
|
作者
Li, Qunwei [1 ]
Zou, Shaofeng [2 ]
Zhong, Wenliang [1 ]
机构
[1] Ant Grp, Hangzhou, Peoples R China
[2] Univ Buffalo State Univ New York, Buffalo, NY USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The first provably efficient algorithm for learning graph neural networks (GNNs) with one hidden layer for node information convolution is provided in this paper. Two types of GNNs are investigated, depending on whether labels are attached to nodes or graphs. A comprehensive framework for designing and analyzing convergence of GNN training algorithms is developed. The algorithm proposed is applicable to a wide range of activation functions including ReLU, Leaky ReLU, Sigmod, Softplus and Swish. It is shown that the proposed algorithm guarantees a linear convergence rate to the underlying true parameters of GNNs. For both types of GNNs, sample complexity in terms of the number of nodes or the number of graphs is characterized. The impact of feature dimension and GNN structure on the convergence rate is also theoretically characterized. Numerical experiments are further provided to validate our theoretical analysis.
引用
收藏
页码:8438 / 8446
页数:9
相关论文
共 50 条
  • [21] Learning Overparameterized Neural Networks via Stochastic Gradient Descent on Structured Data
    Li, Yuanzhi
    Liang, Yingyu
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [22] LEARNING SHALLOW NEURAL NETWORKS VIA PROVABLE GRADIENT DESCENT WITH RANDOM INITIALIZATION
    Xia, Shuhao
    Shi, Yuanming
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 5616 - 5620
  • [23] Crossprop: Learning Representations by Stochastic Meta-Gradient Descent in Neural Networks
    Veeriah, Vivek
    Zhang, Shangtong
    Sutton, Richard S.
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2017, PT I, 2017, 10534 : 445 - 459
  • [24] Applying Gradient Descent in Convolutional Neural Networks
    Cui, Nan
    [J]. 2ND INTERNATIONAL CONFERENCE ON MACHINE VISION AND INFORMATION TECHNOLOGY (CMVIT 2018), 2018, 1004
  • [25] Stochastic Diagonal Approximate Greatest Descent in Neural Networks
    Tan, Hong Hui
    Lim, King Hann
    Harno, Hendra Gunawan
    [J]. 2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 1895 - 1898
  • [26] Scaling Graph Neural Networks with Approximate PageRank
    Bojchevski, Aleksandar
    Klicpera, Johannes
    Perozzi, Bryan
    Kapoor, Amol
    Blais, Martin
    Rozemberczki, Benedek
    Lukasik, Michal
    Guennemann, Stephan
    [J]. KDD '20: PROCEEDINGS OF THE 26TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2020, : 2464 - 2473
  • [27] Fractional-order gradient descent learning of BP neural networks with Caputo derivative
    Wang, Jian
    Wen, Yanqing
    Gou, Yida
    Ye, Zhenyun
    Chen, Hua
    [J]. NEURAL NETWORKS, 2017, 89 : 19 - 30
  • [28] Optimization of learning process for Fourier series neural networks using gradient descent algorithm
    Halawa, Krzysztof
    [J]. PRZEGLAD ELEKTROTECHNICZNY, 2008, 84 (06): : 128 - 130
  • [29] Impact of Mathematical Norms on Convergence of Gradient Descent Algorithms for Deep Neural Networks Learning
    Cai, Linzhe
    Yu, Xinghuo
    Li, Chaojie
    Eberhard, Andrew
    Lien Thuy Nguyen
    Chuong Thai Doan
    [J]. AI 2022: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, 13728 : 131 - 144
  • [30] Gradient descent learning algorithm for hierarchical neural networks: A case study in industrial quality
    Baratta, D
    Diotalevi, F
    Valle, M
    Caviglia, DD
    [J]. ENGINEERING APPLICATIONS OF BIO-INSPIRED ARTIFICIAL NEURAL NETWORKS, VOL II, 1999, 1607 : 578 - 587