Overfitting measurement of convolutional neural networks using trained network weights

被引:3
|
作者
Watanabe, Satoru [1 ]
Yamana, Hayato [2 ]
机构
[1] Waseda Univ, Dept Comp Sci & Commun Engn, Shinjuku Ku, 3-4-1 Okubo, Tokyo 1698555, Japan
[2] Waseda Univ, Fac Sci & Engn, Shinjuku Ku, 3-4-1 Okubo, Tokyo 1698555, Japan
关键词
Convolutional neural network; Overfitting; Persistent homology; Topological data analysis; PERSISTENT HOMOLOGY;
D O I
10.1007/s41060-022-00332-1
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Overfitting reduces the generalizability of convolutional neural networks (CNNs). Overfitting is generally detected by comparing the accuracies and losses of the training and validation data, where the validation data are formed from a portion of the training data; however, detection methods are ineffective for pretrained networks distributed without the training data. Thus, in this paper, we propose a method to detect overfitting of CNNs using the trained network weights inspired by the dropout technique. The dropout technique has been employed to prevent CNNs from overfitting, where the neurons in the CNNs are invalidated randomly during their training. It has been hypothesized that this technique prevents CNNs from overfitting by restraining the co-adaptations among neurons, and this hypothesis implies that the overfitting of CNNs results from co-adaptations among neurons and can be detected by investigating the inner representation of CNNs. The proposed persistent homology-based overfitting measure (PHOM) method constructs clique complexes in CNNs using the trained network weights, and the one-dimensional persistent homology investigates co-adaptations among neurons. In addition, we enhance PHOM to normalized PHOM (NPHOM) to mitigate fluctuation in PHOM caused by the difference in network structures. We applied the proposed methods to convolutional neural networks trained for the classification problems on the CIFAR-10, street view house number, Tiny ImageNet, and CIFAR-100 datasets. Experimental results demonstrate that PHOM and NPHOM can indicate the degree of overfitting of CNNs, which suggests that these methods enable us to filter overfitted CNNs without requiring the training data.
引用
收藏
页码:261 / 278
页数:18
相关论文
共 50 条
  • [1] Overfitting measurement of convolutional neural networks using trained network weights
    Satoru Watanabe
    Hayato Yamana
    International Journal of Data Science and Analytics, 2022, 14 : 261 - 278
  • [2] Overfitting by PSO Trained Feedforward Neural Networks
    van Wyk, Andrich B.
    Engelbrecht, Andries P.
    2010 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2010,
  • [3] Overfitting Measurement of Deep Neural Networks Using No Data
    Watanabe, Satoru
    Yamana, Hayato
    2021 IEEE 8TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (DSAA), 2021,
  • [4] Reducing Overfitting in Deep Convolutional Neural Networks Using Redundancy Regularizer
    Wu, Bingzhe
    Liu, Zhichao
    Yuan, Zhihang
    Sun, Guangyu
    Wu, Charles
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, PT II, 2017, 10614 : 49 - 55
  • [5] Exponential Discretization of Weights of Neural Network Connections in Pre-Trained Neural Networks
    Malsagov, M. Yu
    Khayrov, E. M.
    Pushkareva, M. M.
    Karandashev, I. M.
    OPTICAL MEMORY AND NEURAL NETWORKS, 2019, 28 (04) : 262 - 270
  • [6] Exponential Discretization of Weights of Neural Network Connections in Pre-Trained Neural Networks
    M. Yu. Malsagov
    E. M. Khayrov
    M. M. Pushkareva
    I. M. Karandashev
    Optical Memory and Neural Networks, 2019, 28 : 262 - 270
  • [7] Measuring Overfitting in Convolutional Neural Networks using Adversarial Perturbations and Label Noise
    Pavlitskaya, Svetlana
    Oswald, Joel
    Zollner, J. Marius
    2022 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2022, : 1551 - 1559
  • [8] Convolutional Neural Networks with Fixed Weights
    Folsom, Tyler C.
    VISAPP: PROCEEDINGS OF THE 16TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS - VOL. 5: VISAPP, 2021, : 516 - 523
  • [9] Cropout: A General Mechanism for Reducing Overfitting on Convolutional Neural Networks
    Hou, Wenbo
    Wang, Wenhai
    Liu, Ruo-Ze
    Lu, Tong
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [10] The transition module: a method for preventing overfitting in convolutional neural networks
    Akbar, S.
    Peikari, M.
    Salama, S.
    Nofech-Mozes, S.
    Martel, A. L.
    COMPUTER METHODS IN BIOMECHANICS AND BIOMEDICAL ENGINEERING-IMAGING AND VISUALIZATION, 2019, 7 (03): : 260 - 265