The normalized risk-averting error criterion for avoiding nonglobal local minima in training neural networks

被引:4
|
作者
Lo, James Ting-Ho [1 ]
Gui, Yichuan [2 ]
Peng, Yun [2 ]
机构
[1] Univ Maryland, Dept Math & Stat, Baltimore, MD 21250 USA
[2] Univ Maryland, Dept Comp Sci & Elect Engn, Baltimore, MD 21250 USA
基金
美国国家科学基金会;
关键词
Neural network; Training; Convexification; Risk-averting error; Global optimization; Local minimum;
D O I
10.1016/j.neucom.2013.11.056
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The convexification method for data fitting is capable of avoiding nonglobal local minima, but suffers from two shortcomings: the risk-averting error (RAE) criterion grows exponentially as its risk-sensitivity index A increases, and the existing method of determining A is often not effective. To eliminate these shortcomings, the normalized RAE (NRAE) is herein proposed. As NRAE is a monotone increasing function of RAE, the region without a nonglobal local minimum of NRAE expands as does that of RAE. However, NRAE does not grow unboundedly as does RAE. The performances of training with NRAE at a fixed A are reported. Over a large range of the risk-sensitivity index, such training has a high rate of achieving a global or near global minimum starting with different initial weight vectors of the neural network under training. It is observed that at a large A, the landscape of the NRAE is rather flat, which slows down the training to a halt. This observation motivates the development of the NRAE-MSE method that exploits the large region of an NRAE without a nonglobal local minimum and takes excursions from time to time for training with the standard mean squared error (MSE) to zero into a global or near global minimum. A number of examples of approximating functions that involve fine features or under-sampled segments are used to test the method. Numerical experiments show that the NRAE-MSE training method has a success rate of 100% in all the testing trials for each example, all starting with randomly selected initial weights. The method is also applied to classifying numerals in the well-known MNIST dataset. The new training method outperforms other methods reported in the literature under the same operating conditions. (C) 2014 Elsevier BV. All rights reserved.
引用
收藏
页码:3 / 12
页数:10
相关论文
共 9 条
  • [1] Adaptive Normalized Risk-Averting Training for Deep Neural Networks
    Wang, Zhiguang
    Oates, Tim
    Lo, James
    THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 2201 - 2207
  • [2] A Pairwise Algorithm for Training Multilayer Perceptrons with the Normalized Risk-Averting Error Criterion
    Gui, Yichuan
    Lo, James Ting-Ho
    Peng, Yun
    PROCEEDINGS OF THE 2014 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2014, : 344 - 351
  • [3] Risk-averting criteria for training neural networks
    Lo, JTH
    8TH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING, VOLS 1-3, PROCEEDING, 2001, : 476 - 481
  • [4] Distributed Robust Training of Multilayer Neural Netwroks Using Normalized Risk-Averting Error
    Ninomiya, Hiroshi
    2014 IEEE SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE, COGNITIVE ALGORITHMS, MIND, AND BRAIN (CCMB), 2014, : 134 - 140
  • [5] Avoiding local minima in feedforward neural networks by simultaneous learning
    Atakulreka, Akarachai
    Sutivong, Daricha
    AI 2007: ADVANCES IN ARTIFICIAL INTELLIGENCE, PROCEEDINGS, 2007, 4830 : 100 - +
  • [6] Training Neural Networks with Local Error Signals
    Nokland, Arild
    Eidnes, Lars H.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [7] A normalized adaptive training of recurrent neural networks with augmented error gradient
    Wu Yilei
    Song Qing
    Liu Sheng
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2008, 19 (02): : 351 - 356
  • [8] A Free From Local Minima Algorithm for Training Regressive MLP Neural Networks
    Montisci, Augusto
    ADVANCES IN ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING, 2024, 4 (01): : 2103 - 2112
  • [9] A Systematic Algorithm to Escape from Local Minima in Training Feed-Forward Neural Networks
    Cheung, Chi-Chung
    Xu, Sean Shensheng
    Ng, Sin-Chun
    2016 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2016, : 396 - 402