Automatic model selection for fully connected neural networks

被引:10
|
作者
Laredo D. [1 ]
Ma S.F. [2 ]
Leylaz G. [2 ]
Schütze O. [1 ]
Sun J.-Q. [2 ]
机构
[1] Department of Computer Science, CINVESTAV, Mexico City
[2] Department of Mechanical Engineering, University of California, Merced, 95343, CA
关键词
Artificial neural networks; Distributed computing; Evolutionary algorithms; Hyperparameter tuning; Model selection;
D O I
10.1007/s40435-020-00708-w
中图分类号
学科分类号
摘要
Neural networks and deep learning are changing the way that artificial intelligence is being done. Efficiently choosing a suitable network architecture and fine tuning its hyper-parameters for a specific dataset is a time-consuming task given the staggering number of possible alternatives. In this paper, we address the problem of model selection by means of a fully automated framework for efficiently selecting a neural network model for a selected task, whether it is classification or regression. The algorithm, named Automatic Model Selection, is a modified micro-genetic algorithm that automatically and efficiently finds the most suitable fully connected neural network model for a given dataset. The main contributions of this method are: a simple, list based encoding for neural networks, which will be used as the genotype in our evolutionary algorithm, novel crossover and mutation operators, the introduction of a fitness function that considers the accuracy of the neural network and its complexity, and a method to measure the similarity between two neural networks. AMS is evaluated on two different datasets. By comparing some models obtained with AMS to state-of-the-art models for each dataset we show that AMS can automatically find efficient neural network models. Furthermore, AMS is computationally efficient and can make use of distributed computing paradigms to further boost its performance. © 2020, Springer-Verlag GmbH Germany, part of Springer Nature.
引用
收藏
页码:1063 / 1079
页数:16
相关论文
共 50 条
  • [1] Spectrum Analysis for Fully Connected Neural Networks
    Jia, Bojun
    Zhang, Yanjun
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (12) : 10091 - 10104
  • [2] On the Learnability of Fully-connected Neural Networks
    Zhang, Yuchen
    Lee, Jason D.
    Wainwright, Martin J.
    Jordan, Michael I.
    [J]. ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 54, 2017, 54 : 83 - 91
  • [3] Data Symmetries and Learning in Fully Connected Neural Networks
    Anselmi, Fabio
    Manzoni, Luca
    D'onofrio, Alberto
    Rodriguez, Alex
    Caravagna, Giulio
    Bortolussi, Luca
    Cairoli, Francesca
    [J]. IEEE ACCESS, 2023, 11 : 47282 - 47290
  • [4] A homotopy training algorithm for fully connected neural networks
    Chen, Qipin
    Hao, Wenrui
    [J]. PROCEEDINGS OF THE ROYAL SOCIETY A-MATHEMATICAL PHYSICAL AND ENGINEERING SCIENCES, 2019, 475 (2231):
  • [5] A NOVEL LAYERWISE PRUNING METHOD FOR MODEL REDUCTION OF FULLY CONNECTED DEEP NEURAL NETWORKS
    Mauch, Lukas
    Yang, Bin
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 2382 - 2386
  • [6] The automatic model selection and variable kernel width for RBF neural networks
    Zhou, Peng
    Li, Dehua
    Wu, Hong
    Cheng, Feng
    [J]. NEUROCOMPUTING, 2011, 74 (17) : 3628 - 3637
  • [7] EQUIVALENCE OF APPROXIMATION BY CONVOLUTIONAL NEURAL NETWORKS AND FULLY-CONNECTED NETWORKS
    Petersen, Philipp
    Voigtlaender, Felix
    [J]. PROCEEDINGS OF THE AMERICAN MATHEMATICAL SOCIETY, 2020, 148 (04) : 1567 - 1581
  • [8] VLSI FULLY CONNECTED NEURAL NETWORKS FOR THE IMPLEMENTATION OF OTHER TOPOLOGIES
    CARRABINA, J
    LISA, F
    AVELLANA, N
    PEREZVICENTE, CJ
    VALDERRAMA, E
    [J]. LECTURE NOTES IN COMPUTER SCIENCE, 1991, 540 : 277 - 284
  • [9] Training Fully Connected Neural Networks is ∃R-Complete
    Bertschinger, Daniel
    Hertrich, Christoph
    Jungeblut, Paul
    Miltzow, Tillmann
    Weber, Simon
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [10] Constructive algorithm for fully connected cascade feedforward neural networks
    Qiao, Junfei
    Li, Fanjun
    Han, Honggui
    Li, Wenjing
    [J]. NEUROCOMPUTING, 2016, 182 : 154 - 164