Training Strategies for Convolutional Neural Networks with Transformed Input

被引:0
|
作者
Khandani, Masoumeh Kalantari [1 ]
Mikhael, Wasfy B. [1 ]
机构
[1] Univ Cent Florida, Dept Elect & Comp Engn, Orlando, FL 32816 USA
关键词
image classification; Convolutional neural networks; DCT; DWT; domain transforms;
D O I
10.1109/MWSCAS47672.2021.9531913
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Convolutional Neural Network (CNN) are now considered the main tool for image classification. However, most networks studied for classification are large and have extensive computing and storage requirements. Their training time is also usually very long. Such costly computational and storage requirements cannot be met in many applications with simple devices such as small processors or Internet of things (IoTs) devices. Therefore, reducing the size of networks and input sizes become necessary. However, such reductions are not easy and may reduce the classification performance. We examine how domain transforms under different training strategies can be used for efficient size reduction and improvement of classification accuracy. In this paper, we consider networks with under 220K learnable parameters, as opposed to millions in deeper networks. We show that by representing the input to a CNN using appropriately selected domain transforms, such as discrete wavelet transforms (DWT) or discrete cosine transform (DCT), it is possible to efficiently improve the performance of size-reduced networks. For example, DWT proves to be very effective when significant size reduction is needed (improving the result by up to 9%). It is also shown that by tuning training strategies such as the number of epochs and mini batch size, the performance can be further improved by up to 4% under fixed training time.
引用
收藏
页码:1058 / 1061
页数:4
相关论文
共 50 条
  • [1] On the Robustness of Convolutional Neural Networks Regarding Transformed Input Images
    Timme, Frederik
    Kerdels, Jochen
    Peters, Gabriele
    PROCEEDINGS OF THE 12TH INTERNATIONAL JOINT CONFERENCE ON COMPUTATIONAL INTELLIGENCE (IJCCI), 2020, : 396 - 403
  • [2] CONVOLUTIONAL NEURAL NETWORKS AND TRAINING STRATEGIES FOR SKIN DETECTION
    Kim, Yoonsik
    Hwang, Insung
    Cho, Nam Ik
    2017 24TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2017, : 3919 - 3923
  • [3] The Impact of Input Image Data Size on The Training Speed of Convolutional Neural Networks
    Lyu, Xinzhou
    2021 3RD INTERNATIONAL CONFERENCE ON MACHINE LEARNING, BIG DATA AND BUSINESS INTELLIGENCE (MLBDBI 2021), 2021, : 654 - 657
  • [4] Lightweight and Elegant Data Reduction Strategies for Training Acceleration of Convolutional Neural Networks
    Demidovskij, Alexander
    Tugaryov, Artyom
    Trutnev, Aleksei
    Kazyulina, Marina
    Salnikov, Igor
    Pavlov, Stanislav
    MATHEMATICS, 2023, 11 (14)
  • [5] Latent Training for Convolutional Neural Networks
    Huang, Zi
    Liu, Qi
    Chen, Zhiyuan
    Zhao, Yuming
    PROCEEDINGS OF 2015 INTERNATIONAL CONFERENCE ON ESTIMATION, DETECTION AND INFORMATION FUSION ICEDIF 2015, 2015, : 55 - 60
  • [6] CONVOLUTIONAL NEURAL NETWORKS AND MULTITASK STRATEGIES FOR SEMANTIC MAPPING OF NATURAL LANGUAGE INPUT TO A STRUCTURED DATABASE
    Korpusik, Mandy
    Glass, James
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 6174 - 6178
  • [7] JOINT TRAINING OF CONVOLUTIONAL AND NON-CONVOLUTIONAL NEURAL NETWORKS
    Soltau, Hagen
    Saon, George
    Sainath, Tara N.
    2014 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2014,
  • [8] Towards Improved Input Masking for Convolutional Neural Networks
    Balasubramanian, Sriram
    Feizi, Soheil
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 1855 - 1865
  • [9] Scaling up the training of Convolutional Neural Networks
    Snir, Marc
    2019 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS (IPDPSW), 2019, : 925 - 925
  • [10] Towards dropout training for convolutional neural networks
    Wu, Haibing
    Gu, Xiaodong
    NEURAL NETWORKS, 2015, 71 : 1 - 10