Power law scaling of test error versus number of training images for deep convolutional neural networks

被引:3
|
作者
Sala, Vittorio [1 ]
机构
[1] iMAGE S Spa, Vittorio Alfieri 64, I-22066 Mariano Comense, CO, Italy
关键词
Deep learning; neural networks; image classification; Resnet50; MNIST; computer vision; test error; power law;
D O I
10.1117/12.2525811
中图分类号
TP7 [遥感技术];
学科分类号
081102 ; 0816 ; 081602 ; 083002 ; 1404 ;
摘要
The highest accuracy in image classification is of utmost importance for the industrial application of algorithms based on convolutional neural networks. Empirically, it is sometimes possible to improve accuracy by increasing the size of the training set. In this work, the scaling of the test accuracy versus the size of the training set was studied for different networks. First, a network with a depth of few layers was initialized with random parameters and trained on subsets of images of variable size sampled from the MNIST dataset of handwritten digits and MNIST fashion dataset of clothes and accessories. The scaling of the accuracy versus the size of the training set may be described as the sum of two components: a power law and an offset independent on the size of the training set. Exponent of the power law appears to be the same in both dataset and independent on seeds, initial weights and number of convolutional filters. Then, the scaling of the accuracy versus the size of training set has been evaluated on a dataset of pictures of paintings, sacred icons and sculptures with the goal to correctly classify unknown pictures. The networks chosen are the ones implemented in the machine vision library Halcon 18.11, including two convolutional neural networks with unknown topology and Resnet50, pretrained on industrial images. The scaling of the accuracy versus the size of the training set seems to be compatible with the power law scaling observed on the few layers network trained on MNIST.
引用
收藏
页数:5
相关论文
共 50 条
  • [1] Scaling up the training of Convolutional Neural Networks
    Snir, Marc
    [J]. 2019 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS (IPDPSW), 2019, : 925 - 925
  • [2] An empirical study of the test error versus training error in Artificial Neural Networks
    Dias, Fernando Morgado
    Antunes, Ana
    [J]. PROCEEDINGS OF THE 9TH WSEAS INTERNATIONAL CONFERENCE ON AUTOMATION AND INFORMATION, 2008, : 266 - +
  • [3] On Smart Gaze Based Annotation of Histopathology Images for Training of Deep Convolutional Neural Networks
    Mariam, Komal
    Afzal, Osama Mohammed
    Hussain, Wajahat
    Javed, Muhammad Umar
    Kiyani, Amber
    Rajpoot, Nasir
    Khurram, Syed Ali
    Khan, Hassan Aqeel
    [J]. IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2022, 26 (07) : 3025 - 3036
  • [4] ON CLASSIFICATION OF DISTORTED IMAGES WITH DEEP CONVOLUTIONAL NEURAL NETWORKS
    Zhou, Yiren
    Song, Sibo
    Cheung, Ngai-Man
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 1213 - 1217
  • [5] Deep Convolutional Neural Networks for Fire Detection in Images
    Sharma, Jivitesh
    Granmo, Ole-Christoffer
    Goodwin, Morten
    Fidje, Jahn Thomas
    [J]. ENGINEERING APPLICATIONS OF NEURAL NETWORKS, EANN 2017, 2017, 744 : 183 - 193
  • [6] Efficient Incremental Training for Deep Convolutional Neural Networks
    Tao, Yudong
    Tu, Yuexuan
    Shyu, Mei-Ling
    [J]. 2019 2ND IEEE CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL (MIPR 2019), 2019, : 286 - 291
  • [7] Training Deep Convolutional Neural Networks to Play Go
    Clark, Christopher
    Storkey, Amos
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 37, 2015, 37 : 1766 - 1774
  • [8] Training Deep Convolutional Neural Networks with Active Learning for Exudate Classification in Eye Fundus Images
    Otalora, Sebastian
    Perdomo, Oscar
    Gonzalez, Fabio
    Mueller, Henning
    [J]. INTRAVASCULAR IMAGING AND COMPUTER ASSISTED STENTING, AND LARGE-SCALE ANNOTATION OF BIOMEDICAL DATA AND EXPERT LABEL SYNTHESIS, 2017, 10552 : 146 - 154
  • [9] Full error analysis for the training of deep neural networks
    Beck, Christian
    Jentzen, Arnulf
    Kuckuck, Benno
    [J]. INFINITE DIMENSIONAL ANALYSIS QUANTUM PROBABILITY AND RELATED TOPICS, 2022, 25 (02)
  • [10] Enabling Efficient Training of Convolutional Neural Networks for Histopathology Images
    Alali, Mohammed H.
    Roohi, Arman
    Deogun, Jitender S.
    [J]. IMAGE ANALYSIS AND PROCESSING, ICIAP 2022 WORKSHOPS, PT I, 2022, 13373 : 533 - 544