Automatic model selection for fully connected neural networks

被引:10
|
作者
Laredo D. [1 ]
Ma S.F. [2 ]
Leylaz G. [2 ]
Schütze O. [1 ]
Sun J.-Q. [2 ]
机构
[1] Department of Computer Science, CINVESTAV, Mexico City
[2] Department of Mechanical Engineering, University of California, Merced, 95343, CA
关键词
Artificial neural networks; Distributed computing; Evolutionary algorithms; Hyperparameter tuning; Model selection;
D O I
10.1007/s40435-020-00708-w
中图分类号
学科分类号
摘要
Neural networks and deep learning are changing the way that artificial intelligence is being done. Efficiently choosing a suitable network architecture and fine tuning its hyper-parameters for a specific dataset is a time-consuming task given the staggering number of possible alternatives. In this paper, we address the problem of model selection by means of a fully automated framework for efficiently selecting a neural network model for a selected task, whether it is classification or regression. The algorithm, named Automatic Model Selection, is a modified micro-genetic algorithm that automatically and efficiently finds the most suitable fully connected neural network model for a given dataset. The main contributions of this method are: a simple, list based encoding for neural networks, which will be used as the genotype in our evolutionary algorithm, novel crossover and mutation operators, the introduction of a fitness function that considers the accuracy of the neural network and its complexity, and a method to measure the similarity between two neural networks. AMS is evaluated on two different datasets. By comparing some models obtained with AMS to state-of-the-art models for each dataset we show that AMS can automatically find efficient neural network models. Furthermore, AMS is computationally efficient and can make use of distributed computing paradigms to further boost its performance. © 2020, Springer-Verlag GmbH Germany, part of Springer Nature.
引用
收藏
页码:1063 / 1079
页数:16
相关论文
共 50 条
  • [31] A novel structured sparse fully connected layer in convolutional neural networks
    Matsumura, Naoki
    Ito, Yasuaki
    Nakano, Koji
    Kasagi, Akihiko
    Tabaru, Tsuguchika
    [J]. CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2023, 35 (11):
  • [32] Parallel dynamics of fully connected Q-Ising neural networks
    Bolle, D
    Jongen, G
    Shim, GM
    [J]. JOURNAL OF STATISTICAL PHYSICS, 1998, 91 (1-2) : 125 - 153
  • [33] Thermodynamics of fully connected Blume-Emery-Griffiths neural networks
    Bollé, D
    Verbeiren, T
    [J]. JOURNAL OF PHYSICS A-MATHEMATICAL AND GENERAL, 2003, 36 (02): : 295 - 305
  • [34] Parallel Dynamics of Fully Connected Q-Ising Neural Networks
    D. Bollé
    G. Jongen
    G. M. Shim
    [J]. Journal of Statistical Physics, 1998, 91 : 125 - 153
  • [35] Depth Degeneracy in Neural Networks: Vanishing Angles in Fully Connected ReLU Networks on Initialization
    Jakub, Cameron
    Nica, Mihai
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2024, 25 : 1 - 45
  • [36] Bayesian tensorized neural networks with automatic rank selection
    Hawkins, Cole
    Zhang, Zheng
    [J]. NEUROCOMPUTING, 2021, 453 : 172 - 180
  • [37] Automatic Selection for the Beta Basis Function Neural Networks
    Dhahri, Habib
    Alimi, Adel
    [J]. NATURE INSPIRED COOPERATIVE STRATEGIES FOR OPTIMIZATION (NICSO 2007), 2008, 129 : 461 - 474
  • [38] Finding the Capacity of Fuzzy Neural Networks (FNNs) via Its Equivalent Fully Connected Neural Networks (FFNNs)
    Wang, Jing
    Wang, Chi-Hsu
    Chen, C. L. Philip
    [J]. IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS (FUZZ 2011), 2011, : 2193 - 2198
  • [39] Modified model selection in neural networks
    Xu, L.P.
    Jiang, H.
    Zhang, Y.H.
    [J]. Jisuanji Gongcheng/Computer Engineering, 2001, 27 (02):
  • [40] Fully automatic alpha matte extraction using artificial neural networks
    Roberto Rosas-Romero
    Omar Lopez-Rincon
    Oleg Starostenko
    [J]. Neural Computing and Applications, 2020, 32 : 6843 - 6855