Improved Generalization in Recurrent Neural Networks Using the Tangent Plane Algorithm

被引:0
|
作者
May, P. [1 ]
Zhou, E. [2 ]
Lee, C. W. [2 ]
机构
[1] K Coll, Brook St, Tonbridge, Kent, England
[2] Univ Bolton, Acad Grp, Appl Engn & Sci, Bolton, England
关键词
real time recurrent learning; tangent plane; generalization; weight elimination; temporal pattern recognition; non-linear process control;
D O I
暂无
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
The tangent plane algorithm for real time recurrent learning (TPA-RTRL) is an effective online training method for fully recurrent neural networks. TPA-RTRL uses the method of approaching tangent planes to accelerate the learning processes. Compared to the original gradient descent real time recurrent learning algorithm (GD-RTRL) it is very fast and avoids problems like local minima of the search space. However, the TPA-RTRL algorithm actively encourages the formation of large weight values that can be harmful to generalization. This paper presents a new TPA-RTRL variant that encourages small weight values to decay to zero by using a weight elimination procedure built into the geometry of the algorithm. Experimental results show that the new algorithm gives good generalization over a range of network sizes whilst retaining the fast convergence speed of the TPA-RTRL algorithm.
引用
收藏
页码:118 / 126
页数:9
相关论文
共 50 条
  • [41] An Improved Segmentation of Online English Handwritten Text Using Recurrent Neural Networks
    Cuong Tuan Nguyen
    Nakagawa, Masaki
    PROCEEDINGS 3RD IAPR ASIAN CONFERENCE ON PATTERN RECOGNITION ACPR 2015, 2015, : 176 - 180
  • [42] Optimum design of structures by an improved genetic algorithm using neural networks
    Salajegheh, E
    Gholizadeh, S
    ADVANCES IN ENGINEERING SOFTWARE, 2005, 36 (11-12) : 757 - 767
  • [43] Optimum design of structures by an improved genetic algorithm using neural networks
    Salajegheh, E.
    Gholizadeh, S.
    Proceedings of The Seventh International Conference on the Application of Artificial Intelligence to Civil and Structural Engineering, 2003, : 107 - 108
  • [44] A study on generalization ability of 3-layer recurrent neural networks
    Ninomiya, H
    Sasaki, A
    PROCEEDING OF THE 2002 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-3, 2002, : 1063 - 1068
  • [45] Neural state space alignment for magnitude generalization in humans and recurrent networks
    Sheahan, Hannah
    Luyckx, Fabrice
    Nelli, Stephanie
    Teupe, Clemens
    Summerfield, Christopher
    NEURON, 2021, 109 (07) : 1214 - 1226.e8
  • [46] Training recurrent neural networks by using parallel recursive prediction error algorithm
    Chen, DQ
    Chan, LW
    ICONIP'98: THE FIFTH INTERNATIONAL CONFERENCE ON NEURAL INFORMATION PROCESSING JOINTLY WITH JNNS'98: THE 1998 ANNUAL CONFERENCE OF THE JAPANESE NEURAL NETWORK SOCIETY - PROCEEDINGS, VOLS 1-3, 1998, : 1393 - 1396
  • [47] Nonlinear system identification using genetic algorithm based recurrent neural networks
    Zhu, Yu-Qing
    Xie, Wen-Fang
    Yao, Tie
    2006 CANADIAN CONFERENCE ON ELECTRICAL AND COMPUTER ENGINEERING, VOLS 1-5, 2006, : 1530 - +
  • [48] An on-line learning algorithm for recurrent neural networks using variational methods
    Oh, WG
    Suh, BS
    40TH MIDWEST SYMPOSIUM ON CIRCUITS AND SYSTEMS, VOLS 1 AND 2, 1998, : 659 - 662
  • [49] A distributed genetic algorithm improving the generalization behavior of neural networks
    Branke, J
    Kohlmorgen, U
    Schmeck, H
    MACHINE LEARNING: ECML-95, 1995, 912 : 107 - 121
  • [50] An Improved Time Feedforward Connections Recurrent Neural Networks
    Wang, Jin
    Zou, Yongsong
    Lim, Se -Jung
    INTELLIGENT AUTOMATION AND SOFT COMPUTING, 2023, 36 (03): : 2743 - 2755