AN ANALYSIS OF PREMATURE SATURATION IN BACK-PROPAGATION LEARNING

被引:71
|
作者
LEE, Y
OH, SH
KIM, MW
机构
[1] Electronics and Telecommunications Research Institute, Daejeon
关键词
PREMATURE SATURATION; BACK PROPAGATION ALGORITHM; 1ST EPOCH; MULTILAYER PERCEPTRON;
D O I
10.1016/S0893-6080(05)80116-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The back propagation (BP) algorithm is widely used for finding optimum weights of multilayer neural networks in many pattern recognition applications. However, the critical drawbacks of the algorithm are its slow learning speed and convergence to local minima. One of the major reasons for these drawbacks is the ''premature saturation '' which is a phenomenon that the error of the neural network stays significantly high constant for some period of time during learning. It is known to be caused by an inappropriate set of initial weights. In this paper, the probability of premature saturation at the beginning epoch of learning procedure in the BP algorithm has been derived in terms of the maximum value of initial weights, the number of nodes in each layer, and the maximum slope of the sigmoidal activation function; it has been verified by the Monte Carlo simulation. Using this result, the premature saturation can be avoided with proper initial weight settings.
引用
收藏
页码:719 / 728
页数:10
相关论文
共 50 条
  • [41] Merging back-propagation and Hebbian learning rules for robust classifications
    Jeong, DG
    Lee, SY
    NEURAL NETWORKS, 1996, 9 (07) : 1213 - 1222
  • [42] Back-Propagation Learning in Deep Spike-By-Spike Networks
    Rotermund, David
    Pawelzik, Klaus R.
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2019, 13
  • [43] A complement to the back-propagation algorithm: An upper bound for the learning rate
    Cerqueira, JJF
    Palhares, AGB
    Madrid, MK
    IJCNN 2000: PROCEEDINGS OF THE IEEE-INNS-ENNS INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOL IV, 2000, : 517 - 522
  • [44] A parallel formulation of back-propagation learning on distributed memory multiprocessors
    Mahapatra, S
    Mahapatra, RN
    Chatterji, BN
    PARALLEL COMPUTING, 1997, 22 (12) : 1661 - 1675
  • [45] Adaptive back-propagation in on-line learning of multilayer networks
    West, AHL
    Saad, D
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 8: PROCEEDINGS OF THE 1995 CONFERENCE, 1996, 8 : 323 - 329
  • [46] Convergence of the generalized back-propagation algorithm with constant learning rates
    Ng, SC
    Leung, SH
    Luk, A
    IEEE WORLD CONGRESS ON COMPUTATIONAL INTELLIGENCE, 1998, : 1090 - 1094
  • [47] FEATURE CONSTRUCTION FOR BACK-PROPAGATION
    PIRAMUTHU, S
    LECTURE NOTES IN COMPUTER SCIENCE, 1991, 496 : 264 - 268
  • [48] On the Local Hessian in Back-propagation
    Zhang, Huishuai
    Chen, Wei
    Liu, Tie-Yan
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [49] Localized back-propagation network
    Kongzhi yu Juece Control Decis, 2 (152):
  • [50] A Diffferential Adaptive Learning Rate Method for Back-Propagation Neural Networks
    Iranmanesh, Saeid
    NN'09: PROCEEDINGS OF THE 10TH WSEAS INTERNATIONAL CONFERENCE ON NEURAL NETWORKS, 2009, : 30 - 34