On the generalization ability of recurrent networks

被引:0
|
作者
Hammer, B [1 ]
机构
[1] Univ Osnabruck, Dept Math Comp Sci, D-49069 Osnabruck, Germany
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The generalization ability of discrete time partially recurrent networks is examined. It is well known that the VC dimension of recurrent networks is infinite in most interesting cases and hence the standard VC analysis cannot be applied directly. We find guarantees for specific situations where the transition function forms a contraction or the probability of long inputs is restricted. For the general case, we derive posterior bounds which take the input data into account. They are obtained via a generalization of the luckiness framework to the agnostic setting. The general formalism allows to focus on reppresentative parts of the data as well as more general situations such as long term prediction.
引用
收藏
页码:731 / 736
页数:6
相关论文
共 50 条
  • [41] Neural state space alignment for magnitude generalization in humans and recurrent networks
    Sheahan, Hannah
    Luyckx, Fabrice
    Nelli, Stephanie
    Teupe, Clemens
    Summerfield, Christopher
    NEURON, 2021, 109 (07) : 1214 - 1226.e8
  • [42] Meta-Learning Related Tasks with Recurrent Networks: Optimization and Generalization
    Nguyen, Thy
    Younger, A. Steven
    Redd, Emmett
    Obafemi-Ajayi, Tayo
    2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [43] Synthesis of cascade recurrent neural networks using feedforward generalization properties
    Shaaban, KM
    Schalkoff, RJ
    INFORMATION SCIENCES, 1998, 108 (1-4) : 207 - 217
  • [44] Predicting the generalization ability of neural networks resembling the nearest-neighbor algorithm
    Muselli, M
    IJCNN 2000: PROCEEDINGS OF THE IEEE-INNS-ENNS INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOL I, 2000, : 27 - 32
  • [45] Improving generalization ability of neural networks ensemble with multi-task learning
    State Key Laboratory for Novel Software Technology, Nanjing University, Nanjing 210093, China
    不详
    J. Comput. Inf. Syst., 2006, 4 (1235-1240):
  • [46] Improvement of generalization ability for identifying dynamic systems by using universal learning networks
    Kim, SH
    Hirasawa, K
    Hu, JL
    IJCNN'01: INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-4, PROCEEDINGS, 2001, : 1203 - 1208
  • [47] Study and application of a class of neural networks model whih better generalization ability
    Wang, YC
    Wu, HX
    Geng, CF
    PROCEEDINGS OF THE 4TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-4, 2002, : 2016 - 2020
  • [48] An Approach to Pruning the Structure of Convolutional Neural Networks without Loss of Generalization Ability
    Chen, Chaoxiang
    Kroshchanka, Aliaksandr
    Golovko, Vladimir
    Golovko, Olha
    PATTERN RECOGNITION AND IMAGE ANALYSIS, 2024, 34 (02) : 258 - 265
  • [49] Improvement of generalization ability for identifying dynamical systems by using universal learning networks
    Hirasawa, K
    Kim, S
    Hu, JL
    Murata, J
    Han, M
    Jin, CZ
    NEURAL NETWORKS, 2001, 14 (10) : 1389 - 1404
  • [50] Generalization without Systematicity: On the Compositional Skills of Sequence-to-Sequence Recurrent Networks
    Lake, Brenden
    Baroni, Marco
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80