Learning Graph Neural Networks with Approximate Gradient Descent

被引:0
|
作者
Li, Qunwei [1 ]
Zou, Shaofeng [2 ]
Zhong, Wenliang [1 ]
机构
[1] Ant Grp, Hangzhou, Peoples R China
[2] Univ Buffalo State Univ New York, Buffalo, NY USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The first provably efficient algorithm for learning graph neural networks (GNNs) with one hidden layer for node information convolution is provided in this paper. Two types of GNNs are investigated, depending on whether labels are attached to nodes or graphs. A comprehensive framework for designing and analyzing convergence of GNN training algorithms is developed. The algorithm proposed is applicable to a wide range of activation functions including ReLU, Leaky ReLU, Sigmod, Softplus and Swish. It is shown that the proposed algorithm guarantees a linear convergence rate to the underlying true parameters of GNNs. For both types of GNNs, sample complexity in terms of the number of nodes or the number of graphs is characterized. The impact of feature dimension and GNN structure on the convergence rate is also theoretically characterized. Numerical experiments are further provided to validate our theoretical analysis.
引用
收藏
页码:8438 / 8446
页数:9
相关论文
共 50 条
  • [41] A Caputo-Type Fractional-Order Gradient Descent Learning of BP Neural Networks
    Yang, Guoling
    Zhang, Bingjie
    Sang, Zhaoyang
    Wang, Jian
    Chen, Hua
    [J]. ADVANCES IN NEURAL NETWORKS, PT I, 2017, 10261 : 547 - 554
  • [42] Superpolynomial Lower Bounds for Learning One-Layer Neural Networks using Gradient Descent
    Goel, Surbhi
    Gollakota, Aravind
    Jin, Zhihan
    Karmalkar, Sushrut
    Klivans, Adam
    [J]. 25TH AMERICAS CONFERENCE ON INFORMATION SYSTEMS (AMCIS 2019), 2019,
  • [43] Improving the capacity of complex-valued neural networks with a modified gradient descent learning rule
    Lee, DL
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 2001, 12 (02): : 439 - 443
  • [44] A supervised multi-spike learning algorithm based on gradient descent for spiking neural networks
    Xu, Yan
    Zeng, Xiaoqin
    Han, Lixin
    Yang, Jing
    [J]. NEURAL NETWORKS, 2013, 43 : 99 - 113
  • [45] Superpolynomial Lower Bounds for Learning One-Layer Neural Networks using Gradient Descent
    Goel, Surbhi
    Gollakota, Aravind
    Jin, Zhihan
    Karmalkar, Sushrut
    Klivans, Adam
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [46] An automatic learning rate decay strategy for stochastic gradient descent optimization methods in neural networks
    Wang, Kang
    Dou, Yong
    Sun, Tao
    Qiao, Peng
    Wen, Dong
    [J]. INTERNATIONAL JOURNAL OF INTELLIGENT SYSTEMS, 2022, 37 (10) : 7334 - 7355
  • [47] Is Learning in Biological Neural Networks Based on Stochastic Gradient Descent? An Analysis Using Stochastic Processes
    Christensen, Soeren
    Kallsen, Jan
    [J]. NEURAL COMPUTATION, 2024, 36 (07) : 1424 - 1432
  • [48] One-Pass Online Learning Based on Gradient Descent for Multilayer Spiking Neural Networks
    Lin, Xianghong
    Hu, Tiandou
    Wang, Xiangwen
    [J]. IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2023, 15 (01) : 16 - 31
  • [49] Fast gradient descent algorithm for image classification with neural networks
    El Mouatasim, Abdelkrim
    [J]. SIGNAL IMAGE AND VIDEO PROCESSING, 2020, 14 (08) : 1565 - 1572
  • [50] Convergence of Gradient Descent Algorithm for Diagonal Recurrent Neural Networks
    Xu, Dongpo
    Li, Zhengxue
    Wu, Wei
    Ding, Xiaoshuai
    Qu, Di
    [J]. 2007 SECOND INTERNATIONAL CONFERENCE ON BIO-INSPIRED COMPUTING: THEORIES AND APPLICATIONS, 2007, : 29 - 31