Convergence Analysis of PSO for Hyper-Parameter Selection in Deep Neural Networks

被引:5
|
作者
Nalepa, Jakub [1 ,2 ]
Lorenzo, Pablo Ribalta [1 ]
机构
[1] Future Proc, Gliwice, Poland
[2] Silesian Tech Univ, Gliwice, Poland
关键词
Convergence analysis; PSO; Hyper; parameter selection; DNNs;
D O I
10.1007/978-3-319-69835-9_27
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep Neural Networks (DNNs) have gained enormous research attention since they consistently outperform other state-of-the-art methods in a plethora of machine learning tasks. However, their performance strongly depends on the DNN hyper-parameters which are commonly tuned by experienced practitioners. Recently, we introduced Particle Swarm Optimization (PSO) and parallel PSO techniques to automate this process. In this work, we theoretically and experimentally investigate the convergence capabilities of these algorithms. The experiments were performed for several DNN architectures (both gradually augmented and hand-crafted by a human) using two challenging multi-class benchmark datasets-MNIST and CIFAR-10.
引用
收藏
页码:284 / 295
页数:12
相关论文
共 50 条
  • [1] Particle Swarm Optimization for Hyper-Parameter Selection in Deep Neural Networks
    Lorenzo, Pablo Ribalta
    Nalepa, Jakub
    Kawulok, Michal
    Sanchez Ramos, Luciano
    Ranilla Pastor, Jose
    PROCEEDINGS OF THE 2017 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE (GECCO'17), 2017, : 481 - 488
  • [2] Hyper-Parameter Selection in Deep Neural Networks Using Parallel Particle Swarm Optimization
    Lorenzo, Pablo Ribalta
    Nalepa, Jakub
    Sanchez Ramos, Luciano
    Ranilla Pastor, Jose
    PROCEEDINGS OF THE 2017 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION (GECCO'17 COMPANION), 2017, : 1864 - 1871
  • [3] On hyper-parameter selection for guaranteed convergence of RMSProp
    Liu, Jinlan
    Xu, Dongpo
    Zhang, Huisheng
    Mandic, Danilo
    COGNITIVE NEURODYNAMICS, 2022, 18 (6) : 3227 - 3237
  • [5] Hyper-Parameter Selection in Convolutional Neural Networks Using Microcanonical Optimization Algorithm
    Gulcu, Ayla
    Kus, Zeki
    IEEE ACCESS, 2020, 8 : 52528 - 52540
  • [6] HYPER-PARAMETER OPTIMIZATION OF DEEP CONVOLUTIONAL NETWORKS FOR OBJECT RECOGNITION
    Talathi, Sachin S.
    2015 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2015, : 3982 - 3986
  • [7] USING METAHEURISTICS FOR HYPER-PARAMETER OPTIMIZATION OF CONVOLUTIONAL NEURAL NETWORKS
    Bibaeva, Victoria
    2018 IEEE 28TH INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2018,
  • [8] Neural Networks Designing Neural Networks: Multi-Objective Hyper-Parameter Optimization
    Smithson, Sean C.
    Yang, Guang
    Gross, Warren J.
    Meyer, Brett H.
    2016 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER-AIDED DESIGN (ICCAD), 2016,
  • [9] Sensitivity Analysis for Deep Learning: Ranking Hyper-parameter Influence
    Taylor, Rhian
    Ojha, Varun
    Martino, Ivan
    Nicosia, Giuseppe
    2021 IEEE 33RD INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI 2021), 2021, : 512 - 516
  • [10] Deep neural network hyper-parameter tuning through twofold genetic approach
    Kumar, Puneet
    Batra, Shalini
    Raman, Balasubramanian
    SOFT COMPUTING, 2021, 25 (13) : 8747 - 8771