Automatically Avoiding Overfitting in Deep Neural Networks by Using Hyper-Parameters Optimization Methods

被引:1
|
作者
Kadhim, Zahraa Saddi [1 ]
Abdullah, Hasanen S. [1 ]
Ghathwan, Khalil I. [1 ]
机构
[1] Univ Technol Baghdad, Dept Comp Sci, Baghdad, Iraq
关键词
deep learning; hyper-parameters optimization; regularization; overfitting; SEARCH;
D O I
10.3991/ijoe.v19i05.38153
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
Overfitting is one issue that deep learning faces in particular. It leads to highly accurate classification results, but they are fraudulent. As a result, if the overfitting problem is not fully resolved, systems that rely on prediction or recognition and are sensitive to accuracy will produce untrustworthy results. All prior suggestions helped to lessen this issue but fell short of eliminating it entirely while maintaining crucial data. This paper proposes a novel approach to guarantee the preservation of critical data while eliminating overfitting completely. Numeric and image datasets are employed in two types of networks: convolutional and deep neural networks. Following the usage of three regular-ization techniques (L1, L2, and dropout), apply two optimization algorithms (Bayesian and random search), allowing them to select the hyperparameters automatically, with regularization techniques being one of the hyperparameters that are automatically selected. The obtained results, in addition to completely eliminating the overfitting issue, showed that the accuracy of the image data was 97.82% and 90.72% when using Bayesian and random search techniques, respec-tively, and was 95.3% and 96.5% when using the same algorithms with a numeric dataset.
引用
收藏
页码:146 / 162
页数:17
相关论文
共 50 条
  • [21] Using an Evolutionary Algorithm to Optimize the Hyper-parameters of a Cascading Neural Network
    Vos, Angus
    Plested, Jo
    NEURAL INFORMATION PROCESSING (ICONIP 2019), PT IV, 2019, 1142 : 758 - 765
  • [22] Hyper-parameters Tuning of Artificial Neural Networks: An Application in the Field of Recommender Systems
    Stergiopoulos, Vaios
    Vassilakopoulos, Michael
    Tousidou, Eleni
    Corral, Antonio
    NEW TRENDS IN DATABASE AND INFORMATION SYSTEMS, ADBIS 2022, 2022, 1652 : 266 - 276
  • [23] On the choice of hyper-parameters of artificial neural networks for stabilized finite element schemes
    Subodh M. Joshi
    Thivin Anandh
    Bhanu Teja
    Sashikumaar Ganesan
    International Journal of Advances in Engineering Sciences and Applied Mathematics, 2021, 13 : 278 - 297
  • [24] Hyperspectral Image Classification with Deep CNN Using an Enhanced Elephant Herding Optimization for Updating Hyper-Parameters
    Munishamaiaha, Kavitha
    Kannan, Senthil Kumar
    Venkatesan, DhilipKumar
    Jasinski, Michal
    Novak, Filip
    Gono, Radomir
    Leonowicz, Zbigniew
    ELECTRONICS, 2023, 12 (05)
  • [25] A Study of Features and Deep Neural Network Architectures and Hyper-Parameters for Domestic Audio Classification
    Copiaco, Abigail
    Ritz, Christian
    Abdulaziz, Nidhal
    Fasciani, Stefano
    APPLIED SCIENCES-BASEL, 2021, 11 (11):
  • [26] Extending MLP ANN hyper-parameters Optimization by using Genetic Algorithm
    Itano, Fernando
    de Abreu de Sousa, Miguel Angelo
    Del-Moral-Hernandez, Emilio
    2018 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2018,
  • [27] Effect of hyper-parameters on the performance of ConvLSTM based deep neural network in crop classification
    Durrani, Awab Ur Rashid
    Minallah, Nasru
    Aziz, Najam
    Frnda, Jaroslav
    Khan, Waleed
    Nedoma, Jan
    PLOS ONE, 2023, 18 (02):
  • [28] Optimizing hyper-parameters of neural networks with swarm intelligence: A novel framework for credit scoring
    Zhang, Runchi
    Qiu, Zhiyi
    PLOS ONE, 2020, 15 (06):
  • [29] Deep Recurrent Electricity Theft Detection in AMI Networks with Random Tuning of Hyper-parameters
    Nabil, Mahmoud
    Ismail, Muhammad
    Mahmoud, Mohamed
    Shahin, Mostafa
    Qaraqe, Khalid
    Serpedin, Erchin
    2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 740 - 745
  • [30] AUTOMATED OPTIMIZATION OF DECODER HYPER-PARAMETERS FOR ONLINE LVCSR
    Chandrashekaran, Akshay
    Lane, Ian
    2016 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY (SLT 2016), 2016, : 454 - 460