ANALYSIS OF GRADIENT DESCENT LEARNING ALGORITHMS FOR MULTILAYER FEEDFORWARD NEURAL NETWORKS

被引:12
|
作者
GUO, H
GELFAND, SB
机构
[1] School of Electrical Engineering, Purdue University, West Lafayette
来源
关键词
D O I
10.1109/31.85630
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
We investigate certain dynamical properties of gradient-type learning algorithms as they apply to multilayer feedforward neural networks. These properties are more related to the multilayer structure of the net than to the particular threshold units at the nodes. The analysis explains the empirical observation that the weight sequence generated by backpropagation and related stochastic gradient algorithms exhibits a long term dependence on the initial choice of weights, and also a continued growth and/or drift even long after the outputs have converged. The analysis is carried out in two steps. First a simplified deterministic algorithm is derived using a describing function-type approach. Next, an analysis of the simplified algorithm is performed by considering an associated ordinary differential equation (ODE). Some numerical examples are given to illustrate the analysis. There has been almost no analysis of the dynamical behavior of backpropagation and related algorithms for the training of multilayer nets; this paper represents a first step in that direction.
引用
收藏
页码:883 / 894
页数:12
相关论文
共 50 条
  • [1] Analysis of natural gradient descent for multilayer neural networks
    Rattray, M
    Saad, D
    [J]. PHYSICAL REVIEW E, 1999, 59 (04): : 4523 - 4532
  • [2] Dynamics of on-line gradient descent learning for multilayer neural networks
    Saad, D
    Solla, SA
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 8: PROCEEDINGS OF THE 1995 CONFERENCE, 1996, 8 : 302 - 308
  • [3] Learning curves for stochastic gradient descent in linear feedforward networks
    Werfel, J
    Xie, XH
    Seung, HS
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 16, 2004, 16 : 1197 - 1204
  • [4] Learning curves for stochastic gradient descent in linear feedforward networks
    Werfel, J
    Xie, XH
    Seung, HS
    [J]. NEURAL COMPUTATION, 2005, 17 (12) : 2699 - 2718
  • [5] Fast learning algorithms for feedforward neural networks
    Jiang, MH
    Gielen, G
    Zhang, B
    Luo, ZS
    [J]. APPLIED INTELLIGENCE, 2003, 18 (01) : 37 - 54
  • [6] Diffusion learning algorithms for feedforward neural networks
    Skorohod B.A.
    [J]. Cybernetics and Systems Analysis, 2013, 49 (3) : 334 - 346
  • [7] Fast Learning Algorithms for Feedforward Neural Networks
    Minghu Jiang
    Georges Gielen
    Bo Zhang
    Zhensheng Luo
    [J]. Applied Intelligence, 2003, 18 : 37 - 54
  • [8] LEARNING ALGORITHMS FOR MULTILAYER NEURAL NETWORKS
    AVEDYAN, ED
    [J]. AUTOMATION AND REMOTE CONTROL, 1995, 56 (04) : 541 - 551
  • [9] A fast learning strategy for multilayer feedforward neural networks
    Chen, Huawei
    Zhong, Hualan
    Yuan, Haiying
    Jin, Fan
    [J]. WCICA 2006: SIXTH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-12, CONFERENCE PROCEEDINGS, 2006, : 3019 - +
  • [10] Impact of Mathematical Norms on Convergence of Gradient Descent Algorithms for Deep Neural Networks Learning
    Cai, Linzhe
    Yu, Xinghuo
    Li, Chaojie
    Eberhard, Andrew
    Lien Thuy Nguyen
    Chuong Thai Doan
    [J]. AI 2022: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, 13728 : 131 - 144