Hyper-Parameter Selection in Convolutional Neural Networks Using Microcanonical Optimization Algorithm

被引:27
|
作者
Gulcu, Ayla [1 ]
Kus, Zeki [1 ]
机构
[1] Fatih Sultan Mehmet Univ, Dept Comp Sci, TR-34445 Istanbul, Turkey
关键词
Convolutional neural networks; hyper-parameter optimization; microcanonical optimization; tree Parzen estimator; DROPOUT; SEARCH;
D O I
10.1109/ACCESS.2020.2981141
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The success of Convolutional Neural Networks is highly dependent on the selected architecture and the hyper-parameters. The need for the automatic design of the networks is especially important for complex architectures where the parameter space is so large that trying all possible combinations is computationally infeasible. In this study, Microcanonical Optimization algorithm which is a variant of Simulated Annealing method is used for hyper-parameter optimization and architecture selection for Convolutional Neural Networks. To the best of our knowledge, our study provides a first attempt at applying Microcanonical Optimization for this task. The networks generated by the proposed method is compared to the networks generated by Simulated Annealing method in terms of both accuracy and size using six widely-used image recognition datasets. Moreover, a performance comparison using Tree Parzen Estimator which is a Bayesion optimization-based approach is also presented. It is shown that the proposed method is able to achieve competitive classification results with the state-of-the-art architectures. When the size of the networks is also taken into account, one can see that the networks generated by Microcanonical Optimization method contain far less parameters than the state-of-the-art architectures. Therefore, the proposed method can be preferred for automatically tuning the networks especially in situations where fast training is as important as the accuracy.
引用
收藏
页码:52528 / 52540
页数:13
相关论文
共 50 条
  • [1] USING METAHEURISTICS FOR HYPER-PARAMETER OPTIMIZATION OF CONVOLUTIONAL NEURAL NETWORKS
    Bibaeva, Victoria
    2018 IEEE 28TH INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2018,
  • [2] Hyper-Parameter Selection in Deep Neural Networks Using Parallel Particle Swarm Optimization
    Lorenzo, Pablo Ribalta
    Nalepa, Jakub
    Sanchez Ramos, Luciano
    Ranilla Pastor, Jose
    PROCEEDINGS OF THE 2017 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION (GECCO'17 COMPANION), 2017, : 1864 - 1871
  • [3] Particle Swarm Optimization for Hyper-Parameter Selection in Deep Neural Networks
    Lorenzo, Pablo Ribalta
    Nalepa, Jakub
    Kawulok, Michal
    Sanchez Ramos, Luciano
    Ranilla Pastor, Jose
    PROCEEDINGS OF THE 2017 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE (GECCO'17), 2017, : 481 - 488
  • [4] HYPER-PARAMETER OPTIMIZATION OF DEEP CONVOLUTIONAL NETWORKS FOR OBJECT RECOGNITION
    Talathi, Sachin S.
    2015 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2015, : 3982 - 3986
  • [5] Multi-objective simulated annealing for hyper-parameter optimization in convolutional neural networks
    Gulcu, Ayla
    Kus, Zeki
    PEERJ COMPUTER SCIENCE, 2021, 7 : 2 - 27
  • [6] Classification of Similar Sports Images Using Convolutional Neural Network with Hyper-Parameter Optimization
    Podgorelec, Vili
    Pecnik, Spela
    Vrbancic, Grega
    APPLIED SCIENCES-BASEL, 2020, 10 (23): : 1 - 24
  • [7] Hyper-Parameter Optimization by Using the Genetic Algorithm for Upper Limb Activities Recognition Based on Neural Networks
    Zhang, Junjie
    Sun, Guangmin
    Sun, Yuge
    Dou, Huijing
    Bilal, Anas
    IEEE SENSORS JOURNAL, 2021, 21 (02) : 1877 - 1884
  • [8] Hyper-Parameter Optimization by Using the Genetic Algorithm for Upper Limb Activities Recognition Based on Neural Networks
    Faculty of Information Technology, Beijing University of Technology, Beijing, China
    不详
    IEEE Sensors J., 1600, 2 (1877-1884):
  • [9] Neural Networks Designing Neural Networks: Multi-Objective Hyper-Parameter Optimization
    Smithson, Sean C.
    Yang, Guang
    Gross, Warren J.
    Meyer, Brett H.
    2016 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER-AIDED DESIGN (ICCAD), 2016,
  • [10] Convergence Analysis of PSO for Hyper-Parameter Selection in Deep Neural Networks
    Nalepa, Jakub
    Lorenzo, Pablo Ribalta
    ADVANCES ON P2P, PARALLEL, GRID, CLOUD AND INTERNET COMPUTING (3PGCIC-2017), 2018, 13 : 284 - 295