Tight Bounds on the Smallest Eigenvalue of the Neural Tangent Kernel for Deep ReLU Networks

被引:0
|
作者
Nguyen, Quynh [1 ]
Mondelli, Marco [2 ]
Montufar, Guido [1 ,3 ]
机构
[1] MPI MIS, Leipzig, Germany
[2] IST Austria, Klosterneuburg, Austria
[3] Univ Calif Los Angeles, Los Angeles, CA 90024 USA
基金
欧洲研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A recent line of work has analyzed the theoretical properties of deep neural networks via the Neural Tangent Kernel (NTK). In particular, the smallest eigenvalue of the NTK has been related to the memorization capacity, the global convergence of gradient descent algorithms and the generalization of deep nets. However, existing results either provide bounds in the two-layer setting or assume that the spectrum of the NTK matrices is bounded away from 0 for multi-layer networks. In this paper, we provide tight bounds on the smallest eigenvalue of NTK matrices for deep ReLU nets, both in the limiting case of infinite widths and for finite widths. In the finite-width setting, the network architectures we consider are fairly general: we require the existence of a wide layer with roughly order of N neurons, N being the number of data samples; and the scaling of the remaining layer widths is arbitrary (up to logarithmic factors). To obtain our results, we analyze various quantities of independent interest: we give lower bounds on the smallest singular value of hidden feature matrices, and upper bounds on the Lipschitz constant of input-output feature maps.
引用
收藏
页数:11
相关论文
共 50 条
  • [31] Analyzing Finite Neural Networks: CanWe Trust Neural Tangent Kernel Theory?
    Seleznova, Mariia
    Kutyniok, Gitta
    MATHEMATICAL AND SCIENTIFIC MACHINE LEARNING, VOL 145, 2021, 145 : 868 - 895
  • [32] Quantum-classical hybrid neural networks in the neural tangent kernel regime
    Nakaji, Kouhei
    Tezuka, Hiroyuki
    Yamamoto, Naoki
    QUANTUM SCIENCE AND TECHNOLOGY, 2024, 9 (01)
  • [33] Graph Neural Tangent Kernel: Fusing Graph Neural Networks with Graph Kernels
    Du, Simon S.
    Hou, Kangcheng
    Poczos, Barnabas
    Salakhutdinov, Ruslan
    Wang, Ruosong
    Xu, Keyulu
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [34] Tight Certified Robustness via Min-Max Representations of ReLU Neural Networks
    Anderson, Brendon G.
    Pfrommer, Samuel
    Sojoudi, Somayeh
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 6348 - 6355
  • [35] Disentangling the Predictive Variance of Deep Ensembles through the Neural Tangent Kernel
    Kobayashi, Seijin
    Aceituno, Pau Vilimelis
    von Oswald, Johannes
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [36] On the CVP for the root lattices via folding with deep ReLU neural networks
    Corlay, Vincent
    Boutros, Joseph J.
    Ciblat, Philippe
    Brunel, Loic
    2019 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2019, : 1622 - 1626
  • [37] Deep learning in random neural fields: Numerical experiments via neural tangent kernel
    Watanabe, Kaito
    Sakamoto, Kotaro
    Karakida, Ryo
    Sonoda, Sho
    Amari, Shun-ichi
    NEURAL NETWORKS, 2023, 160 : 148 - 163
  • [38] Error bounds for deep ReLU networks using the Kolmogorov-Arnold superposition theorem
    Montanelli, Hadrien
    Yang, Haizhao
    NEURAL NETWORKS, 2020, 129 : 1 - 6
  • [39] NONPARAMETRIC REGRESSION USING DEEP NEURAL NETWORKS WITH RELU ACTIVATION FUNCTION
    Schmidt-Hieber, Johannes
    ANNALS OF STATISTICS, 2020, 48 (04): : 1875 - 1897
  • [40] Approximation in shift-invariant spaces with deep ReLU neural networks
    Yang, Yunfei
    Li, Zhen
    Wang, Yang
    NEURAL NETWORKS, 2022, 153 : 269 - 281