A recalling-enhanced recurrent neural network: Conjugate gradient learning algorithm and its convergence analysis

被引:50
|
作者
Gao, Tao [1 ,2 ]
Gong, Xiaoling [1 ,2 ]
Zhang, Kai [3 ]
Lin, Feng [2 ,4 ]
Wang, Jian [2 ]
Huang, Tingwen [5 ]
Zurada, Jacek M. [6 ,7 ]
机构
[1] China Univ Petr East China, Coll Control Sci & Engn, Qingdao 266580, Peoples R China
[2] China Univ Petr East China, Coll Sci, Qingdao 266580, Peoples R China
[3] China Univ Petr East China, Sch Petr Engn, Qingdao 266580, Peoples R China
[4] Beijing Sunwad Sci & Technol Ltd, Beijing 100124, Peoples R China
[5] Texas A&M Univ Qatar, Doha 23874, Qatar
[6] Univ Louisville, Elect & Comp Engn, Louisville, KY 40292 USA
[7] Univ Social Sci, Informat Technol Inst, PL-90113 Lodz, Poland
基金
中国国家自然科学基金;
关键词
Recurrent; Neural network; Conjugate gradient; Generalized Armijo search; Monotonicity; Convergence; DETERMINISTIC CONVERGENCE; MINIMIZATION;
D O I
10.1016/j.ins.2020.01.045
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Elman network is a classical recurrent neural network with an internal delay feedback. In this paper, we propose a recalling-enhanced recurrent neural network (RERNN) which has a selective memory property. In addition, an improved conjugate algorithm with generalized Armijo search technique that speeds up the convergence rate is used to train the RERNN model. Further enhancement performance is achieved with adaptive learning coefficients. Finally, we prove weak and strong convergence of the presented algorithm. In other words, as the number of training steps increases, the following has been established for RERNN: (1) the gradient norm of the error function with respect to the weight vectors converges to zero, (2) the weight sequence approaches a fixed optimal point. We have carried out a number of simulations to illustrate and verify the theoretical results that demonstrate the efficiency of the proposed algorithm. (C) 2020 Elsevier Inc. All rights reserved.
引用
收藏
页码:273 / 288
页数:16
相关论文
共 50 条
  • [1] Recalling-enhanced recurrent neural network optimized with wood pecker mating algorithm for brain tumor classification
    Suganthy, M.
    Ashok, S.
    Maheswari, A. Uma
    Subha, T. D.
    [J]. CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2023, 35 (22):
  • [2] Recalling-Enhanced Recurrent Neural Network optimized with Chimp Optimization Algorithm based speech enhancement for hearing aids
    Rai, Rahul R.
    Mathivanan, M.
    [J]. INTELLIGENT DECISION TECHNOLOGIES-NETHERLANDS, 2024, 18 (01): : 123 - 134
  • [3] A conjugate gradient learning algorithm for recurrent neural networks
    Chang, WF
    Mak, MW
    [J]. NEUROCOMPUTING, 1999, 24 (1-3) : 173 - 189
  • [5] IoT-based COVID-19 detection using recalling-enhanced recurrent neural network optimized with golden eagle optimization algorithm
    Karthick, S.
    Gomathi, N.
    [J]. MEDICAL & BIOLOGICAL ENGINEERING & COMPUTING, 2024, 62 (03) : 925 - 940
  • [6] Parameter Conjugate Gradient with Secant Equation Based Elman Neural Network and its Convergence Analysis
    Fan, Qinwei
    Zhang, Zhiwen
    Huang, Xiaodi
    [J]. ADVANCED THEORY AND SIMULATIONS, 2022, 5 (09)
  • [7] Training Bidirectional Recurrent Neural Network Architectures with the Scaled Conjugate Gradient Algorithm
    Agathocleous, Michalis
    Christodoulou, Chris
    Promponas, Vasilis
    Kountouris, Petros
    Vassiliades, Vassilis
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2016, PT I, 2016, 9886 : 123 - 131
  • [8] Power Quality (PQ) Analyses of DG Utilizing Unified Power Quality Conditioner (UPQC) by White Shark Optimizer and Recalling-Enhanced Recurrent Neural Network
    Shravani, Chapala
    Narasimham, R. L.
    Das, G. Tulasi Ram
    [J]. JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2024, 33 (13)
  • [9] Convergence of gradient method for a fully recurrent neural network
    Xu, Dongpo
    Li, Zhengxue
    Wu, Wei
    [J]. SOFT COMPUTING, 2010, 14 (03) : 245 - 250
  • [10] Convergence of gradient method for a fully recurrent neural network
    Dongpo Xu
    Zhengxue Li
    Wei Wu
    [J]. Soft Computing, 2010, 14 : 245 - 250