A normalized adaptive training of recurrent neural networks with augmented error gradient

被引:4
|
作者
Wu Yilei [1 ]
Song Qing [1 ]
Liu Sheng [1 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore 639798, Singapore
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 2008年 / 19卷 / 02期
关键词
adaptive learning rate; augmented error gradient; convergence; normalization;
D O I
10.1109/TNN.2007.908647
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
For training algorithms of recurrent neural networks (RNN), convergent speed and training error are always two contradictory performances. In this letter, we propose a normalized adaptive recurrent learning (NARL) to obtain a tradeoff between transient and steady-state response. An augmented term is added to error gradient to exactly model the derivative of cost function with respect to hidden layer weight. The influence of the induced gain of activation function on training stability is also taken into consideration. Moreover, adaptive learning rate is employed to improve the robustness of the gradient training. Fianlly, computer simulations of a model prediction problem are synthesized to give comparisons between NARL and conventional normalized real-time recurrent learning (N-RTRL).
引用
收藏
页码:351 / 356
页数:6
相关论文
共 50 条
  • [1] Adaptive nonmonotone conjugate gradient training algorithm for recurrent neural networks
    Peng, Chun-Cheng
    Magoulas, George D.
    [J]. 19TH IEEE INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, VOL II, PROCEEDINGS, 2007, : 374 - 381
  • [2] ADVANCED ADAPTIVE NONMONOTONE CONJUGATE GRADIENT TRAINING ALGORITHM FOR RECURRENT NEURAL NETWORKS
    Peng, Chun-Cheng
    Magoulas, George D.
    [J]. INTERNATIONAL JOURNAL ON ARTIFICIAL INTELLIGENCE TOOLS, 2008, 17 (05) : 963 - 984
  • [3] A normalized gradient algorithm for an adaptive recurrent perceptron
    Chambers, JA
    Sherliker, W
    Mandic, DP
    [J]. 2000 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, PROCEEDINGS, VOLS I-VI, 2000, : 396 - 399
  • [4] Adaptive Normalized Risk-Averting Training for Deep Neural Networks
    Wang, Zhiguang
    Oates, Tim
    Lo, James
    [J]. THIRTIETH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2016, : 2201 - 2207
  • [5] BATCH NORMALIZED RECURRENT NEURAL NETWORKS
    Laurent, Cesar
    Pereyra, Gabriel
    Brakel, Philemon
    Zhang, Ying
    Bengio, Yoshua
    [J]. 2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 2657 - 2661
  • [6] GRADIENT EXPLOSION FREE ALGORITHM FOR TRAINING RECURRENT NEURAL NETWORKS
    Hong, Seoyoung
    Jeon, Hyerin
    Lee, Byungjoon
    Min, Chohong
    [J]. JOURNAL OF THE KOREAN SOCIETY FOR INDUSTRIAL AND APPLIED MATHEMATICS, 2020, 24 (04) : 331 - 350
  • [7] Rates of learning in gradient and genetic training of recurrent neural networks
    Riaza, R
    Zufiria, PJ
    [J]. ARTIFICIAL NEURAL NETS AND GENETIC ALGORITHMS, 1999, : 95 - 99
  • [8] Robust Adaptive Gradient-Descent Training Algorithm for Recurrent Neural Networks in Discrete Time Domain
    Song, Qing
    Wu, Yilei
    Soh, Yeng Chai
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS, 2008, 19 (11): : 1841 - 1853
  • [9] An improved recursive prediction error algorithm for training recurrent neural networks
    Li, HR
    Wang, XZ
    Gu, SS
    [J]. PROCEEDINGS OF THE 3RD WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-5, 2000, : 1043 - 1046
  • [10] Gradient-free training of recurrent neural networks using random perturbations
    Fernandez, Jesus Garcia
    Keemink, Sander
    van Gerven, Marcel
    [J]. FRONTIERS IN NEUROSCIENCE, 2024, 18