Evolutionary Optimization of Hyperparameters in Deep Learning Models

被引:0
|
作者
Kim, Jin-Young [1 ]
Cho, Sung-Bae [1 ]
机构
[1] Yonsei Univ, Dept Comp Sci, Seoul, South Korea
关键词
Genetic programming; deep learning; neural networks; activation function; optimization technique; NEURAL-NETWORKS;
D O I
10.1109/cec.2019.8790354
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Recently, deep learning is one of the most popular techniques in artificial intelligence. However, to construct a deep learning model, various components must be set up, including activation functions, optimization methods, a configuration of model structure called hyperparameters. As they affect the performance of deep learning, researchers are working hard to find optimal hyperparameters when solving problems with deep learning. Activation function and optimization technique play a crucial role in the forward and backward processes of model learning, but they are set up in a heuristic way. The previous studies have been conducted to optimize either activation function or optimization technique, while the relationship between them is neglected to search them at the same time. In this paper, we propose a novel method based on genetic programming to simultaneously find the optimal activation functions and optimization techniques. In genetic programming, each individual is composed of two chromosomes, one for the activation function and the other for the optimization technique. To calculate the fitness of one individual, we construct a neural network with the activation function and optimization technique that the individual represents. The deep learning model found through our method has 82.59% and 53.04% of accuracies for the CIFAR-10 and CIFAR-100 datasets, which outperforms the conventional methods. Moreover, we analyze the activation function found and confirm the usefulness of the proposed method.
引用
收藏
页码:831 / 837
页数:7
相关论文
共 50 条
  • [31] Parameters Optimization of Deep Learning Models using Particle Swarm Optimization
    Qolomany, Basheer
    Maabreh, Majdi
    Al-Fuqaha, Ala
    Gupta, Ajay
    Benhaddou, Driss
    2017 13TH INTERNATIONAL WIRELESS COMMUNICATIONS AND MOBILE COMPUTING CONFERENCE (IWCMC), 2017, : 1285 - 1290
  • [32] Optimization on selecting XGBoost hyperparameters using meta-learning
    Lima Marinho, Tiago
    do Nascimento, Diego Carvalho
    Pimentel, Bruno Almeida
    EXPERT SYSTEMS, 2024, 41 (09)
  • [33] The Deep Learning ResNet101 and Ensemble XGBoost Algorithm with Hyperparameters Optimization Accurately Predict the Lung Cancer
    Ahmed, Saghir
    Raza, Basit
    Hussain, Lal
    Aldweesh, Amjad
    Omar, Abdulfattah
    Khan, Mohammad Shahbaz
    Eldin, Elsayed Tag
    Nadim, Muhammad Amin
    APPLIED ARTIFICIAL INTELLIGENCE, 2023, 37 (01)
  • [34] Combining Evolutionary Algorithms and Deep Learning for Hardware/Software Interface Optimization
    Servadei, Lorenzo
    Mosca, Edoardo
    Werner, Michael
    Esen, Volkan
    Wille, Robert
    Ecker, Wolfgang
    2019 ACM/IEEE 1ST WORKSHOP ON MACHINE LEARNING FOR CAD (MLCAD), 2019,
  • [35] Hybridization of Evolutionary Algorithm and Deep Reinforcement Learning for Multiobjective Orienteering Optimization
    Liu, Wei
    Wang, Rui
    Zhang, Tao
    Li, Kaiwen
    Li, Wenhua
    Ishibuchi, Hisao
    Liao, Xiangke
    IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2023, 27 (05) : 1260 - 1274
  • [36] Discrete simulation optimization for tuning machine learning method hyperparameters
    Ramamohan, Varun
    Singhal, Shobhit
    Gupta, Aditya Raj
    Bolia, Nomesh Bhojkumar
    JOURNAL OF SIMULATION, 2024, 18 (05) : 745 - 765
  • [37] The role of hyperparameters in machine learning models and how to tune them
    Arnold, Christian
    Biedebach, Luka
    Kuepfer, Andreas
    Neunhoeffer, Marcel
    POLITICAL SCIENCE RESEARCH AND METHODS, 2024, 12 (04) : 841 - 848
  • [38] Evolutionary computation to explain deep learning models for time series forecasting
    Troncoso-Garcia, A. R.
    Martinez-Ballesteros, M.
    Martinez-Alvarez, F.
    Troncoso, A.
    38TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, SAC 2023, 2023, : 433 - 436
  • [39] Hyperparameters Tuning for Machine Learning Models for Time Series Forecasting
    Peter, Gladilin
    Matskevichus, Maria
    2019 SIXTH INTERNATIONAL CONFERENCE ON SOCIAL NETWORKS ANALYSIS, MANAGEMENT AND SECURITY (SNAMS), 2019, : 328 - 332
  • [40] Integrated Evolutionary Learning: An Artificial Intelligence Approach to Joint Learning of Features and Hyperparameters for Optimized, Explainable Machine Learning
    de Lacy, Nina
    Ramshaw, Michael J.
    Kutz, J. Nathan
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2022, 5