Improved Generalization in Recurrent Neural Networks Using the Tangent Plane Algorithm

被引:0
|
作者
May, P. [1 ]
Zhou, E. [2 ]
Lee, C. W. [2 ]
机构
[1] K Coll, Brook St, Tonbridge, Kent, England
[2] Univ Bolton, Acad Grp, Appl Engn & Sci, Bolton, England
关键词
real time recurrent learning; tangent plane; generalization; weight elimination; temporal pattern recognition; non-linear process control;
D O I
暂无
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
The tangent plane algorithm for real time recurrent learning (TPA-RTRL) is an effective online training method for fully recurrent neural networks. TPA-RTRL uses the method of approaching tangent planes to accelerate the learning processes. Compared to the original gradient descent real time recurrent learning algorithm (GD-RTRL) it is very fast and avoids problems like local minima of the search space. However, the TPA-RTRL algorithm actively encourages the formation of large weight values that can be harmful to generalization. This paper presents a new TPA-RTRL variant that encourages small weight values to decay to zero by using a weight elimination procedure built into the geometry of the algorithm. Experimental results show that the new algorithm gives good generalization over a range of network sizes whilst retaining the fast convergence speed of the TPA-RTRL algorithm.
引用
收藏
页码:118 / 126
页数:9
相关论文
共 50 条
  • [1] PRUNING RECURRENT NEURAL NETWORKS FOR IMPROVED GENERALIZATION PERFORMANCE
    GILES, CL
    OMLIN, CW
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 1994, 5 (05): : 848 - 851
  • [2] Neural Tangent Kernel: Convergence and Generalization in Neural Networks
    Jacot, Arthur
    Gabriel, Franck
    Hongler, Clement
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [3] A learning algorithm for improved recurrent neural networks
    Chen, CH
    Yu, LW
    1997 IEEE INTERNATIONAL CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, 1997, : 2198 - 2202
  • [4] An improved locally linear embedding algorithm by using recurrent neural networks
    Zhang W.
    Journal of Computational and Theoretical Nanoscience, 2016, 13 (01) : 425 - 437
  • [5] Training feedforward neural networks: An algorithm giving improved generalization
    Lee, CW
    NEURAL NETWORKS, 1997, 10 (01) : 61 - 68
  • [6] An improved diagonal recurrent neural networks iterative algorithm
    Hu, HG
    Liu, QA
    Yan, XJ
    Er, LJ
    PROCEEDINGS OF THE FIFTH INTERNATIONAL CONFERENCE ON FLUID POWER TRANSMISSION AND CONTROL (ICFP'2001), 2001, : 524 - 527
  • [7] Neural Tangent Kernel: Convergence and Generalization in Neural Networks (Invited Paper)
    Jacot, Arthur
    Gabriel, Franck
    Hongler, Clement
    STOC '21: PROCEEDINGS OF THE 53RD ANNUAL ACM SIGACT SYMPOSIUM ON THEORY OF COMPUTING, 2021, : 6 - 6
  • [8] On the Provable Generalization of Recurrent Neural Networks
    Wang, Lifu
    Shen, Bo
    Hu, Bo
    Cao, Xing
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [9] An improved blind equalization algorithm based on recurrent neural networks
    Xu, Rong
    Dianzi Keji Daxue Xuebao/Journal of the University of Electronic Science and Technology of China, 2007, 36 (02): : 210 - 212
  • [10] Neural Tangent Generalization Attacks
    Yuan, Chia-Hung
    Wu, Shan-Hung
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139