CURIOUS: Efficient Neural Architecture Search Based on a Performance Predictor and Evolutionary Search

被引:4
|
作者
Hassantabar, Shayan [1 ]
Dai, Xiaoliang [2 ]
Jha, Niraj K. [1 ]
机构
[1] Princeton Univ, Dept Elect & Comp Engn, Princeton, NJ 08544 USA
[2] Facebook Mobile Vis, Campton, CA USA
关键词
Compression; convolutional neural network (CNN); deep learning; dimensionality reduction; long short-term memory (LSTM); neural architecture search (NAS); transformer;
D O I
10.1109/TCAD.2022.3148202
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Neural networks (NNs) have been successfully deployed in various applications of artificial intelligence. However, architectural design of NNs is still a challenging problem. This is due to the need to navigate a search space based on a large number of hyperparameters. This forces the search space of possible architectures to grow exponentially. Using a trial-and-error design approach is very time consuming and leads to suboptimal architectures. In addition, approaches, such as neural architecture search based on reinforcement learning and differentiable gradient-based architecture search, often incur huge computational costs or significant memory requirements. To address these challenges, we propose the CURIOUS NN synthesis methodology. It uses a performance predictor to efficiently navigate the architectural search space with an evolutionary search process. The predictor is built using quasi Monte-Carlo sampling, boosted decision tree regression, and an intelligent iterative sampling method. It is designed to be sample efficient. CURIOUS starts from a base architecture and explores the architectural search space to obtain a variant of the base architecture with the highest performance. This search framework is general and covers all important NN architecture types, e.g., feedforward NNs (FFNNs), convolutional NNs (CNNs), recurrent NNs (RNNs), and transformers. We evaluate the performance of CURIOUS on various datasets and base architectures. Through these experiments, we demonstrate significant performance improvements over the baseline architectures. For theMNIST dataset, our CNN architecture achieves an error rate of 0.66%, with 8.6x fewer parameters compared to the LeNet-5 baseline. For the CIFAR-10 dataset, we use the ResNet architectures and residual networks with Shake-Shake regularization as the baselines. Our synthesized ResNet-18 has a 2.52% accuracy improvement over the original ResNet-18, 1.74% over ResNet-101, and 0.16% over ResNet-1001, while requiring comparable number of parameters and floating-point operations to the original ResNet-18. This result shows that instead of just increasing the number of layers to increase accuracy, an alternative is to use a better NN architecture with a small number of layers. In addition, CURIOUS achieves an error rate of just 2.69% with a variant of the residual architecture with Shake-Shake regularization. We also use the set of optimized hyperparameters found for ResNet-18 on the CIFAR-10 dataset to train and evaluate the model on the ImageNet dataset, and show 3.43% (1.83%) improvement in the top-1 (top-5) error rate compared to the original ResNet-18 model. CURIOUS also obtains the highest accuracy for various other FFNNs that are geared toward edge devices and IoT sensors. In addition, we use CURIOUS to search for deep RNN architectures for the SICK dataset for sentence similarity evaluation. It achieves a mean-squared error of only 0.2060, improving upon the base network performance, without the need to stack multiple long short-term memories. We also use CURIOUS to search for a better NN classifier for the sentiment analysis task on the Stanford sentiment treebank dataset using a pretrained BERT model and again demonstrate improvements in performance.
引用
收藏
页码:4975 / 4990
页数:16
相关论文
共 50 条
  • [41] Fast Evolutionary Neural Architecture Search Based on Bayesian Surrogate Model
    Shi, Rui
    Luo, Jianping
    Liu, Qiqi
    2021 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC 2021), 2021, : 1217 - 1224
  • [42] Evolutionary neural architecture search based on evaluation correction and functional units
    Shang, Ronghua
    Zhu, Songling
    Ren, Jinhong
    Liu, Hangcheng
    Jiao, Licheng
    KNOWLEDGE-BASED SYSTEMS, 2022, 251
  • [43] Constrained Sampling-Based Evolutionary Neural Architecture Search for GANs
    Yang, Yeming
    Zhu, Qingling
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT II, ICIC 2024, 2024, 14863 : 257 - 268
  • [44] Network Intrusion Detection Based on an Efficient Neural Architecture Search
    Lyu, Renjian
    He, Mingshu
    Zhang, Yu
    Jin, Lei
    Wang, Xinlei
    SYMMETRY-BASEL, 2021, 13 (08):
  • [45] A Novel Training Protocol for Performance Predictors of Evolutionary Neural Architecture Search Algorithms
    Sun, Yanan
    Sun, Xian
    Fang, Yuhan
    Yen, Gary G.
    Liu, Yuqiao
    IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2021, 25 (03) : 524 - 536
  • [46] Evolutionary Neural Architecture Search Supporting Approximate Multipliers
    Pinos, Michal
    Mrazek, Vojtech
    Sekanina, Lukas
    GENETIC PROGRAMMING, EUROGP 2021, 2021, 12691 : 82 - 97
  • [47] Evolutionary neural architecture search on transformers for RUL prediction
    Mo, Hyunho
    Iacca, Giovanni
    MATERIALS AND MANUFACTURING PROCESSES, 2023, 38 (15) : 1881 - 1898
  • [48] Comparison of Topologies Generated by Evolutionary Neural Architecture Search
    Yoo, YongSuk
    Park, Manbok
    Park, Kang-Moon
    APPLIED SCIENCES-BASEL, 2023, 13 (09):
  • [49] Evolutionary Neural Architecture Search and Its Applications in Healthcare
    Liu, Xin
    Li, Jie
    Zhao, Jianwei
    Cao, Bin
    Yan, Rongge
    Lyu, Zhihan
    CMES-COMPUTER MODELING IN ENGINEERING & SCIENCES, 2024, 139 (01): : 143 - 185
  • [50] Evolutionary Neural Architecture Search for Facial Expression Recognition
    Deng, Shuchao
    Lv, Zeqiong
    Galvan, Edgar
    Sun, Yanan
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2023, 7 (05): : 1405 - 1419