Hyper-Parameter Selection in Convolutional Neural Networks Using Microcanonical Optimization Algorithm

被引:27
|
作者
Gulcu, Ayla [1 ]
Kus, Zeki [1 ]
机构
[1] Fatih Sultan Mehmet Univ, Dept Comp Sci, TR-34445 Istanbul, Turkey
关键词
Convolutional neural networks; hyper-parameter optimization; microcanonical optimization; tree Parzen estimator; DROPOUT; SEARCH;
D O I
10.1109/ACCESS.2020.2981141
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The success of Convolutional Neural Networks is highly dependent on the selected architecture and the hyper-parameters. The need for the automatic design of the networks is especially important for complex architectures where the parameter space is so large that trying all possible combinations is computationally infeasible. In this study, Microcanonical Optimization algorithm which is a variant of Simulated Annealing method is used for hyper-parameter optimization and architecture selection for Convolutional Neural Networks. To the best of our knowledge, our study provides a first attempt at applying Microcanonical Optimization for this task. The networks generated by the proposed method is compared to the networks generated by Simulated Annealing method in terms of both accuracy and size using six widely-used image recognition datasets. Moreover, a performance comparison using Tree Parzen Estimator which is a Bayesion optimization-based approach is also presented. It is shown that the proposed method is able to achieve competitive classification results with the state-of-the-art architectures. When the size of the networks is also taken into account, one can see that the networks generated by Microcanonical Optimization method contain far less parameters than the state-of-the-art architectures. Therefore, the proposed method can be preferred for automatically tuning the networks especially in situations where fast training is as important as the accuracy.
引用
收藏
页码:52528 / 52540
页数:13
相关论文
共 50 条
  • [21] A New Hyper-Parameter Optimization Method for Power Load Forecast Based on Recurrent Neural Networks
    Li, Yaru
    Zhang, Yulai
    Cai, Yongping
    ALGORITHMS, 2021, 14 (06)
  • [22] Physics-informed recurrent neural networks and hyper-parameter optimization for dynamic process systems
    Asrav, Tuse
    Aydin, Erdal
    COMPUTERS & CHEMICAL ENGINEERING, 2023, 173
  • [23] Random search for hyper-parameter optimization
    Département D'Informatique et de Recherche Opérationnelle, Université de Montréal, Montréal, QC, H3C 3J7, Canada
    J. Mach. Learn. Res., (281-305):
  • [24] Experimental evaluation of stochastic configuration networks: Is SC algorithm inferior to hyper-parameter optimization method?
    Hu, Minghui
    Suganthan, P.N.
    Applied Soft Computing, 2022, 126
  • [25] Random Search for Hyper-Parameter Optimization
    Bergstra, James
    Bengio, Yoshua
    JOURNAL OF MACHINE LEARNING RESEARCH, 2012, 13 : 281 - 305
  • [26] Experimental evaluation of stochastic configuration networks: Is SC algorithm inferior to hyper-parameter optimization method?
    Hu, Minghui
    Suganthan, P. N.
    APPLIED SOFT COMPUTING, 2022, 126
  • [27] Hyper-parameter Optimization for Latent Spaces
    Veloso, Bruno
    Caroprese, Luciano
    Konig, Matthias
    Teixeira, Sonia
    Manco, Giuseppe
    Hoos, Holger H.
    Gama, Joao
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2021: RESEARCH TRACK, PT III, 2021, 12977 : 249 - 264
  • [28] Federated learning with hyper-parameter optimization
    Kundroo, Majid
    Kim, Taehong
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2023, 35 (09)
  • [29] An Enhanced Hyper-Parameter Optimization of a Convolutional Neural Network Model for Leukemia Cancer Diagnosis in a Smart Healthcare System
    Awotunde, Joseph Bamidele
    Imoize, Agbotiname Lucky
    Ayoade, Oluwafisayo Babatope
    Abiodun, Moses Kazeem
    Do, Dinh-Thuan
    Silva, Adao
    Sur, Samarendra Nath
    SENSORS, 2022, 22 (24)
  • [30] Array Nulling Synthesis Based on Hyper-Parameter Optimized Self-Paced Learning Convolutional Neural Networks
    Zhang, Yinghao
    Hu, Haoquan
    Li, Tianyang
    Chen, Bo
    Tian, Jing
    Lei, Shiwen
    IEEE TRANSACTIONS ON ANTENNAS AND PROPAGATION, 2024, 72 (06) : 5325 - 5330