A second-order learning algorithm for multilayer networks based on block Hessian matrix

被引:17
|
作者
Wang, YJ [1 ]
Lin, CT [1 ]
机构
[1] Natl Chiao Tung Univ, Dept Elect & Control Engn, Hsinchu, Taiwan
关键词
multilayer perceptrons; Hessian matrix; forward-backward propagation; Newton's method; least squares estimation;
D O I
10.1016/S0893-6080(98)00091-4
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This article proposes a new second-order learning algorithm for training the multilayer perceptron (MLP) networks. The proposed algorithm is a revised Newton's method. A forward-backward propagation scheme is first proposed for network computation of the Hessian matrix, H, of the output error function of the MLP. A block Hessian matrix, H-b, is then defined to approximate and simplify H. Several lemmas and theorems are proved to uncover the important properties of H and H-b, and verify the good approximation of H-b to H; H-b preserves the major properties of H. The theoretic analysis leads to the development of an efficient way for computing the inverse of H-b recursively. In the proposed second-order learning algorithm, the least squares estimation technique is adopted to further lessen the local minimum problems. The proposed algorithm overcomes not only the drawbacks of the standard backpropagation algorithm (i.e. slow asymptotic convergence rate, bad controllability of convergence accuracy, local minimum problems, and high sensitivity to learning constant), but also the shortcomings of normal Newton's method used on the MLP, such as the lack of network implementation of H, ill representability of the diagonal terms of H, the heavy computation load of the inverse of H, and the requirement of a good initial estimate of the solution (weights). Several example problems are used to demonstrate the efficiency of the proposed learning algorithm. Extensive performance (convergence rate and accuracy) comparisons of the proposed algorithm with other learning schemes (including the standard backpropagation algorithm) are also made. (C) 1998 Elsevier Science Ltd. All rights reserved.
引用
收藏
页码:1607 / 1622
页数:16
相关论文
共 50 条
  • [1] An efficient learning algorithm with second-order convergence for multilayer neural networks
    Ninomiya, H
    Tomita, C
    Asai, H
    PROCEEDINGS OF THE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS 2003, VOLS 1-4, 2003, : 2028 - 2032
  • [2] Second-order observability of matrix-weight-based networks
    Lv, Pengchao
    Huang, Junjie
    Liu, Bo
    Su, Housheng
    Chen, Alatancang
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2023, 360 (04): : 2769 - 2793
  • [3] Annealing based dynamic learning in second-order neural networks
    Milenkovic, S
    Obradovic, Z
    Litovski, V
    ICNN - 1996 IEEE INTERNATIONAL CONFERENCE ON NEURAL NETWORKS, VOLS. 1-4, 1996, : 458 - 463
  • [4] Second-order backpropagation algorithms for a stagewise-partitioned separable Hessian matrix
    Mizutani, E
    Dreyfus, SE
    Demmel, JW
    PROCEEDINGS OF THE INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), VOLS 1-5, 2005, : 1027 - 1032
  • [5] SpikeLM: A second-order supervised learning algorithm for training spiking neural networks
    Wang, Yongji
    Huang, Jian
    DYNAMICS OF CONTINUOUS DISCRETE AND IMPULSIVE SYSTEMS-SERIES A-MATHEMATICAL ANALYSIS, 2006, 13 : 646 - 650
  • [6] Fast second order learning algorithm for feedforward multilayer neural networks and its applications
    Osowski, S
    Bojarczak, P
    Stodolski, M
    NEURAL NETWORKS, 1996, 9 (09) : 1583 - 1596
  • [7] Fast second order learning algorithm for feedforward multilayer neural networks and its applications
    Warsaw University of Technology, Radom, Poland
    不详
    不详
    NEURAL NETW., 9 (1583-1596):
  • [8] Second-order stagewise backpropagation for Hessian-matrix analyses and investigation of negative curvature
    Mizutani, Eiji
    Dreyfus, Stuart E.
    NEURAL NETWORKS, 2008, 21 (2-3) : 193 - 203
  • [9] Second-order learning algorithm with squared penalty term
    Saito, K
    Nakano, R
    NEURAL COMPUTATION, 2000, 12 (03) : 709 - 729
  • [10] Second-order learning algorithm with squared penalty term
    Saito, K
    Nakano, R
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 9: PROCEEDINGS OF THE 1996 CONFERENCE, 1997, 9 : 627 - 633