Dynamics of Deep Neural Networks and Neural Tangent Hierarchy

被引:0
|
作者
Huang, Jiaoyang [1 ]
Yau, Horng-Tzer [2 ]
机构
[1] IAS, Sch Math, Princeton, NJ 08540 USA
[2] Harvard, Math Dept, Cambridge, MA USA
关键词
GAME; GO;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The evolution of a deep neural network trained by the gradient descent in the overparametrization regime can be described by its neural tangent kernel (NTK) (Jacot et al., 2013; Du et al., 2018b;a; Arora et al., 2019b). It was observed (Arora et al., 2019a) that there is a performance gap between the kernel regression using the limiting NTK and the deep neural networks. We study the dynamic of neural networks of finite width and derive an infinite hierarchy of differential equations, the neural tangent hierarchy (NTH). We prove that the NTH hierarchy truncated at the level p >= 2 approximates the dynamic of the NTK up to arbitrary precision under certain conditions on the neural network width and the data set dimension. The assumptions needed for these approximations become weaker as p increases. Finally, NTH can be viewed as higher order extensions of NTK. In particular, the NTH truncated at p = 2 recovers the NTK dynamics.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Neural Tangent Kernel Analysis of Deep Narrow Neural Networks
    Lee, Jongmin
    Choi, Joo Young
    Ryu, Ernest K.
    No, Albert
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [2] On the Neural Tangent Kernel of Deep Networks with Orthogonal Initialization
    Huang, Wei
    Du, Weitao
    Da Xu, Richard Yi
    [J]. PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2577 - 2583
  • [3] Abstraction Hierarchy in Deep Learning Neural Networks
    Ilin, Roman
    Watson, Thomas
    Kozma, Robert
    [J]. 2017 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2017, : 768 - 774
  • [4] Towards an Understanding of Residual Networks Using Neural Tangent Hierarchy (NTH)
    Li, Yuqing
    Luo, Tao
    Yip, Nung Kwan
    [J]. CSIAM TRANSACTIONS ON APPLIED MATHEMATICS, 2022, 3 (04): : 692 - 760
  • [5] Spectral Analysis of the Neural Tangent Kernel for Deep Residual Networks
    Belfer, Yuval
    Geifman, Amnon
    Galun, Meirav
    Basri, Ronen
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2024, 25 : 1 - 49
  • [6] Selection dynamics for deep neural networks
    Liu, Hailiang
    Markowich, Peter
    [J]. JOURNAL OF DIFFERENTIAL EQUATIONS, 2020, 269 (12) : 11540 - 11574
  • [7] Deep neural networks and molecular dynamics
    Car, Roberto
    [J]. ABSTRACTS OF PAPERS OF THE AMERICAN CHEMICAL SOCIETY, 2018, 256
  • [8] "Lossless" Compression of Deep Neural Networks: A High-dimensional Neural Tangent Kernel Approach
    Gu, Lingyu
    Du, Yongqi
    Zhang, Yuan
    Xie, Di
    Pu, Shiliang
    Qiu, Robert C.
    Liao, Zhenyu
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [9] Neural Tangent Kernel: Convergence and Generalization in Neural Networks
    Jacot, Arthur
    Gabriel, Franck
    Hongler, Clement
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [10] Empirical study on tangent loss function for classification with deep neural networks
    Zhang, Xu
    Lu, Wenpeng
    Pan, Yan
    Wu, Hao
    Wang, Rongyao
    Yu, Rui
    [J]. COMPUTERS & ELECTRICAL ENGINEERING, 2021, 90