Convergence Analysis of PSO for Hyper-Parameter Selection in Deep Neural Networks

被引:5
|
作者
Nalepa, Jakub [1 ,2 ]
Lorenzo, Pablo Ribalta [1 ]
机构
[1] Future Proc, Gliwice, Poland
[2] Silesian Tech Univ, Gliwice, Poland
关键词
Convergence analysis; PSO; Hyper; parameter selection; DNNs;
D O I
10.1007/978-3-319-69835-9_27
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep Neural Networks (DNNs) have gained enormous research attention since they consistently outperform other state-of-the-art methods in a plethora of machine learning tasks. However, their performance strongly depends on the DNN hyper-parameters which are commonly tuned by experienced practitioners. Recently, we introduced Particle Swarm Optimization (PSO) and parallel PSO techniques to automate this process. In this work, we theoretically and experimentally investigate the convergence capabilities of these algorithms. The experiments were performed for several DNN architectures (both gradually augmented and hand-crafted by a human) using two challenging multi-class benchmark datasets-MNIST and CIFAR-10.
引用
收藏
页码:284 / 295
页数:12
相关论文
共 50 条
  • [31] LSTM Hyper-Parameter Selection for Malware Detection: Interaction Effects and Hierarchical Selection Approach
    Sewak, Mohit
    Sahay, Sanjay K.
    Rathore, Hemant
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [32] Hyper-parameter Comparison on Convolutional Neural Network for Visual Aerial Localization
    Berhold, J. Mark
    Leishman, Robert C.
    Borghetti, Brett
    Venable, Donald
    PROCEEDINGS OF THE ION 2019 PACIFIC PNT MEETING, 2019, : 875 - 885
  • [33] Hyper-Parameter Optimization by Using the Genetic Algorithm for Upper Limb Activities Recognition Based on Neural Networks
    Faculty of Information Technology, Beijing University of Technology, Beijing, China
    不详
    IEEE Sensors J., 1600, 2 (1877-1884):
  • [34] Learning sparse linear dynamic networks in a hyper-parameter free setting
    Venkitaraman, Arun
    Hjalmarsson, Hakan
    Wahlberg, Bo
    IFAC PAPERSONLINE, 2020, 53 (02): : 82 - 86
  • [35] Hyper-parameter optimization of gradient boosters for flood susceptibility analysis
    Lai, Tuan Anh
    Nguyen, Ngoc-Thach
    Bui, Quang-Thanh
    TRANSACTIONS IN GIS, 2023, 27 (01) : 224 - 238
  • [36] A New Approach Towards the Combined Algorithm Selection and Hyper-parameter Optimization Problem
    Guo, Xin
    van Stein, Bas
    Back, Thomas
    2019 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2019), 2019, : 2042 - 2049
  • [37] A review of automatic selection methods for machine learning algorithms and hyper-parameter values
    Luo, Gang
    NETWORK MODELING AND ANALYSIS IN HEALTH INFORMATICS AND BIOINFORMATICS, 2016, 5 (01):
  • [38] Hyper-parameter Optimization of Convolutional Neural Networks for Classifying COVID-19 X-ray Images
    Vrbacic, Grega
    Pecnik, Spela
    Podgorelec, Vili
    COMPUTER SCIENCE AND INFORMATION SYSTEMS, 2022, 19 (01) : 327 - 352
  • [39] Array Nulling Synthesis Based on Hyper-Parameter Optimized Self-Paced Learning Convolutional Neural Networks
    Zhang, Yinghao
    Hu, Haoquan
    Li, Tianyang
    Chen, Bo
    Tian, Jing
    Lei, Shiwen
    IEEE TRANSACTIONS ON ANTENNAS AND PROPAGATION, 2024, 72 (06) : 5325 - 5330
  • [40] Feature Selection and Support Vector Machine Hyper-parameter Optimisation for Spam Detection
    Diale, Melvin
    Van der Walt, Christiaan
    Celik, Turgay
    Modupe, Abiodun
    2016 PATTERN RECOGNITION ASSOCIATION OF SOUTH AFRICA AND ROBOTICS AND MECHATRONICS INTERNATIONAL CONFERENCE (PRASA-ROBMECH), 2016,