Automated CNN optimization using multi-objective grammatical evolution

被引:0
|
作者
da Silva, Cleber A. C. F. [1 ]
Rosa, Daniel Carneiro [1 ]
Miranda, Pericles B. C. [1 ]
Si, Tapas [2 ]
Cerri, Ricardo [3 ]
Basgalupp, Marcio P. [4 ]
机构
[1] Univ Fed Rural Pernambuco, Recife, PE, Brazil
[2] Univ Engn & Management, Jaipur, Rajasthan, India
[3] Univ Fed Sao Carlos, Sao Carlos, SP, Brazil
[4] Univ Fed Sao Paulo, Sao Jose Dos Campos, SP, Brazil
基金
巴西圣保罗研究基金会;
关键词
Grammatical evolution; Convolutional neural networks; Multi-objective optimization; Automated machine learning;
D O I
10.1016/j.asoc.2023.111124
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Selecting and optimizing Convolutional Neural Networks (CNNs) has become a very complex task given the number of associated optimizable parameters, as well as the fact that the arrangement of the layers present in a CNN directly influences its performance. Several research areas used automation techniques to construct and optimize these architectures, with Grammatical Evolution (GE) being one of the most promising techniques. Although several works proposed solutions to the problem in question, each adopts its own evaluation strategy (e.g., different datasets, evaluation metrics, hardware infrastructure). This divergence makes it difficult to compare the proposed approaches, and consequently, it is not possible to reach safe conclusions about the performance of the solutions. This work proposes an experimental evaluation of several context-free grammars listed in the literature for constructing and optimizing CNNs architectures. In addition, we included four well-known CNNs as baselines: DenseNet169, EfficientNetB1, InceptionV3 and ResNet50V2. We aim to identify the best practices for elaborating grammars and compare their results with consolidated CNNs for image classification problems in the literature. Besides, we assessed all approaches on the same controlled environment (e.g., datasets, evaluation metrics, software and hardware setup) to guarantee fairness in the evaluation process. The experiments were carried out by investigating the performance of the models generated by different grammars in solving image classification problems in three datasets of variable dimensions: CIFAR-10, EuroSAT, and MNIST. The experiments have validated several key findings: (i) the significance of optimizing Convolutional Neural Networks (CNNs); (ii) the potential of grammar-based methods as a promising alternative for this task, yielding CNN models that outperform state-of-the-art CNN architectures while possessing fewer trainable parameters, resulting in reduced computational complexity; (iii) grammars incorporating regularization layers (such as dropout and batch normalization) and those that confine the search space (via parameter constraints on CNNs) consistently produce high-performing models with lower complexity, even after a few generations of the evolutionary process; and (iv) the selection of the grammar for optimization can positively or negatively impact the model generation, depending on the specific task requirements.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Multi-Objective Optimization of Dynamic Memory Managers using Grammatical Evolution
    Manuel Colmenar, J.
    Risco-Martin, Jose L.
    Atienza, David
    Hidalgo, J. Ignacio
    Felipe, C. E. S., II
    [J]. GECCO-2011: PROCEEDINGS OF THE 13TH ANNUAL GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, 2011, : 1819 - 1826
  • [2] A hybrid automated trading system based on multi-objective grammatical evolution
    Contreras, Ivan
    Ignacio Hidalgo, J.
    Nunez-Letamendia, Laura
    [J]. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2017, 32 (03) : 2461 - 2475
  • [3] Improving multi-objective evolutionary algorithms using Grammatical Evolution
    Rodriguez, Amin V. Bernabe
    Alejo-Cerezo, Braulio I.
    Coello, Carlos A. Coello
    [J]. SWARM AND EVOLUTIONARY COMPUTATION, 2024, 84
  • [4] Automated Metamodel/Model Co-evolution Using a Multi-objective Optimization Approach
    Kessentini, Wael
    Sahraoui, Houari
    Wimmer, Manuel
    [J]. MODELLING FOUNDATIONS AND APPLICATIONS, ECMFA 2016, 2016, 9764 : 138 - 155
  • [5] Multi-objective evolution strategy for multimodal multi-objective optimization
    Zhang, Kai
    Chen, Minshi
    Xu, Xin
    Yen, Gary G.
    [J]. APPLIED SOFT COMPUTING, 2021, 101
  • [6] AutoOC: Automated multi-objective design of deep autoencoders and one-class classifiers using grammatical evolution
    Ferreira, Luis
    Cortez, Paulo
    [J]. APPLIED SOFT COMPUTING, 2023, 144
  • [7] Procedural texture evolution using multi-objective optimization
    Brian J. Ross
    Han Zhu
    [J]. New Generation Computing, 2004, 22 : 271 - 293
  • [8] Procedural texture evolution using multi-objective optimization
    Ross, BJ
    Zhu, H
    [J]. NEW GENERATION COMPUTING, 2004, 22 (03) : 271 - 293
  • [9] Automated watershed subdivision for simulations using multi-objective optimization
    Momm, Henrique G.
    Bingner, Ronald L.
    Emilaire, Randall
    Garbrecht, Jurgen
    Wells, Robert R.
    Kuhnle, Roger A.
    [J]. HYDROLOGICAL SCIENCES JOURNAL-JOURNAL DES SCIENCES HYDROLOGIQUES, 2017, 62 (10): : 1564 - 1582
  • [10] Differential evolution for multi-objective optimization
    Babu, BV
    Jehan, MML
    [J]. CEC: 2003 CONGRESS ON EVOLUTIONARY COMPUTATION, VOLS 1-4, PROCEEDINGS, 2003, : 2696 - 2703