Empirical estimation of generalization ability of neural networks

被引:0
|
作者
Sarkar, D. [1 ]
机构
[1] Department of Computer Science, University of Miami, Coral Gables, FL 33124, United States
关键词
Backpropagation - Computer architecture - Learning algorithms - Parameter estimation;
D O I
暂无
中图分类号
学科分类号
摘要
This work concentrates on a novel method for empirical estimation of generalization ability of neural networks. Given a set of training (and testing) data, one can choose a network architecture (number of layers, number of neurons in each layer etc.), an initialization method, and a learning algorithm to obtain a network. One measure of the performance of a trained network is how closely its actual output approximates the desired output for an input that it has never seen before. Current methods provide a number that indicates the estimation of the generalization ability of the network. However, this number provides no further information to understand the contributing factors when the generalization ability is not very good. The method proposed uses a number of parameters to define the generalization ability. A set of the values of these parameters provide an estimate of the generalization ability. In addition, the value of each parameter indicates the contribution of such factors as network architecture, initialization method, training data set, etc. Furthermore, a method has been developed to verify the validity of the estimated values of the parameters.
引用
收藏
页码:3 / 15
相关论文
共 50 条
  • [41] An Empirical Study on Improving the Speed and Generalization of Neural Networks Using a Parallel Circuit Approach
    Kien Tuong Phan
    Maul, Tomas Henrique
    Tuong Thuy Vu
    [J]. INTERNATIONAL JOURNAL OF PARALLEL PROGRAMMING, 2017, 45 (04) : 780 - 796
  • [42] ON THE GENERALIZATION ABILITY OF NEURAL-NETWORK CLASSIFIERS
    MUSAVI, MT
    CHAN, KH
    HUMMELS, DM
    KALANTRI, K
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 1994, 16 (06) : 659 - 663
  • [43] A new method to improve the generalization ability of neural networks: A case study of nuclear mass training
    Zhao, Tianliang
    Zhang, Hongfei
    [J]. NUCLEAR PHYSICS A, 2022, 1021
  • [44] Mining and Injecting Legal Prior Knowledge to Improve the Generalization Ability of Neural Networks in Chinese Judgments
    Chen, Yaying
    Ma, Ji
    Gu, Nanfei
    Wang, Xintong
    He, Minghua
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT V, 2023, 14258 : 350 - 362
  • [45] GENERALIZATION ABILITY OF OPTIMAL CLUSTER SEPARATION NETWORKS
    WENDEMUTH, A
    [J]. JOURNAL OF PHYSICS A-MATHEMATICAL AND GENERAL, 1994, 27 (07): : 2325 - 2333
  • [46] Leveraging the Generalization Ability of Deep Convolutional Neural Networks for Improving Classifiers for Color Fundus Photographs
    Son, Jaemin
    Kim, Jaeyoung
    Kong, Seo Taek
    Jung, Kyu-Hwan
    [J]. APPLIED SCIENCES-BASEL, 2021, 11 (02): : 1 - 10
  • [47] Fault Diagnosis Using Neural Networks for Parallel Shaft Gearboxes and Discussion on Its Generalization Ability
    Liu, Zhanchi
    Sun, Heqing
    [J]. PROCEEDINGS OF 2020 IEEE 9TH DATA DRIVEN CONTROL AND LEARNING SYSTEMS CONFERENCE (DDCLS'20), 2020, : 1288 - 1292
  • [48] REIN: A Robust Training Method for Enhancing Generalization Ability of Neural Networks in Autonomous Driving Systems
    Yu, Fuxun
    Liu, Chenchen
    Chen, Xian
    [J]. 24TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC 2019), 2019, : 456 - 461
  • [49] Achieving Robust Generalization for Wireless Channel Estimation Neural Networks by Designed Training Data
    Luan, Dianxin
    Thompson, John
    [J]. ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 3462 - 3467
  • [50] Wavelet Neural Networks Generalization Improvement
    Skhiri, Mohamed Zine El Abidine
    Chtourou, Mohamed
    [J]. 2013 10TH INTERNATIONAL MULTI-CONFERENCE ON SYSTEMS, SIGNALS & DEVICES (SSD), 2013,