Empirical estimation of generalization ability of neural networks

被引:0
|
作者
Sarkar, D. [1 ]
机构
[1] Department of Computer Science, University of Miami, Coral Gables, FL 33124, United States
关键词
Backpropagation - Computer architecture - Learning algorithms - Parameter estimation;
D O I
暂无
中图分类号
学科分类号
摘要
This work concentrates on a novel method for empirical estimation of generalization ability of neural networks. Given a set of training (and testing) data, one can choose a network architecture (number of layers, number of neurons in each layer etc.), an initialization method, and a learning algorithm to obtain a network. One measure of the performance of a trained network is how closely its actual output approximates the desired output for an input that it has never seen before. Current methods provide a number that indicates the estimation of the generalization ability of the network. However, this number provides no further information to understand the contributing factors when the generalization ability is not very good. The method proposed uses a number of parameters to define the generalization ability. A set of the values of these parameters provide an estimate of the generalization ability. In addition, the value of each parameter indicates the contribution of such factors as network architecture, initialization method, training data set, etc. Furthermore, a method has been developed to verify the validity of the estimated values of the parameters.
引用
收藏
页码:3 / 15
相关论文
共 50 条
  • [1] Empirical estimation of generalization ability of neural networks
    Sarkar, D
    [J]. APPLICATIONS AND SCIENCE OF ARTIFICIAL NEURAL NETWORKS II, 1996, 2760 : 54 - 60
  • [2] ON THE ABILITY OF NEURAL NETWORKS TO PERFORM GENERALIZATION BY INDUCTION
    ANSHELEVICH, VV
    AMIRIKIAN, BR
    LUKASHIN, AV
    FRANKKAMENETSKII, MD
    [J]. BIOLOGICAL CYBERNETICS, 1989, 61 (02) : 125 - 128
  • [3] Extract Generalization Ability from Convolutional Neural Networks
    Wu, Huan
    Wu, JunMin
    Ding, Jie
    [J]. 2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018, : 729 - 734
  • [4] Ensemble Approach for Improving Generalization Ability of Neural Networks
    Ahmed, Shaib
    Razib, Md. Razibul Islam
    Alam, Mohammed Shamsul
    Alam, Mohammad Shafiul
    Huda, Mohammad Nurul
    [J]. 2013 INTERNATIONAL CONFERENCE ON INFORMATICS, ELECTRONICS & VISION (ICIEV), 2013,
  • [5] Improving the generalization ability of neural networks by interval arithmetic
    Ishibuchi, H
    Nii, M
    [J]. 1998 SECOND INTERNATIONAL CONFERENCE ON KNOWLEDGE-BASED INTELLIGENT ELECTRONIC SYSTEMS, KES'98 PROCEEDINGS, VOL 1, 1998, : 231 - 236
  • [6] Bayesian model comparison versus generalization ability of neural networks
    Gomari, M
    Järvi, T
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL I AND II, 1999, : 537 - 541
  • [8] A Novel Ensemble Approach for Improving Generalization Ability of Neural Networks
    Lu, Lei
    Zeng, Xiaoqin
    Wu, Shengli
    Zhong, Shuiming
    [J]. INTELLIGENT DATA ENGINEERING AND AUTOMATED LEARNING - IDEAL 2008, 2008, 5326 : 164 - +
  • [9] Quantitative analysis of the generalization ability of deep feedforward neural networks
    Yang, Yanli
    Li, Chenxia
    [J]. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2021, 40 (03) : 4867 - 4876
  • [10] Fuzzification of input vectors for improving the generalization ability of neural networks
    Ishibuchi, H
    Nii, M
    [J]. 1998 IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS AT THE IEEE WORLD CONGRESS ON COMPUTATIONAL INTELLIGENCE - PROCEEDINGS, VOL 1-2, 1998, : 1153 - 1158