CURIOUS: Efficient Neural Architecture Search Based on a Performance Predictor and Evolutionary Search

被引:4
|
作者
Hassantabar, Shayan [1 ]
Dai, Xiaoliang [2 ]
Jha, Niraj K. [1 ]
机构
[1] Princeton Univ, Dept Elect & Comp Engn, Princeton, NJ 08544 USA
[2] Facebook Mobile Vis, Campton, CA USA
关键词
Compression; convolutional neural network (CNN); deep learning; dimensionality reduction; long short-term memory (LSTM); neural architecture search (NAS); transformer;
D O I
10.1109/TCAD.2022.3148202
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Neural networks (NNs) have been successfully deployed in various applications of artificial intelligence. However, architectural design of NNs is still a challenging problem. This is due to the need to navigate a search space based on a large number of hyperparameters. This forces the search space of possible architectures to grow exponentially. Using a trial-and-error design approach is very time consuming and leads to suboptimal architectures. In addition, approaches, such as neural architecture search based on reinforcement learning and differentiable gradient-based architecture search, often incur huge computational costs or significant memory requirements. To address these challenges, we propose the CURIOUS NN synthesis methodology. It uses a performance predictor to efficiently navigate the architectural search space with an evolutionary search process. The predictor is built using quasi Monte-Carlo sampling, boosted decision tree regression, and an intelligent iterative sampling method. It is designed to be sample efficient. CURIOUS starts from a base architecture and explores the architectural search space to obtain a variant of the base architecture with the highest performance. This search framework is general and covers all important NN architecture types, e.g., feedforward NNs (FFNNs), convolutional NNs (CNNs), recurrent NNs (RNNs), and transformers. We evaluate the performance of CURIOUS on various datasets and base architectures. Through these experiments, we demonstrate significant performance improvements over the baseline architectures. For theMNIST dataset, our CNN architecture achieves an error rate of 0.66%, with 8.6x fewer parameters compared to the LeNet-5 baseline. For the CIFAR-10 dataset, we use the ResNet architectures and residual networks with Shake-Shake regularization as the baselines. Our synthesized ResNet-18 has a 2.52% accuracy improvement over the original ResNet-18, 1.74% over ResNet-101, and 0.16% over ResNet-1001, while requiring comparable number of parameters and floating-point operations to the original ResNet-18. This result shows that instead of just increasing the number of layers to increase accuracy, an alternative is to use a better NN architecture with a small number of layers. In addition, CURIOUS achieves an error rate of just 2.69% with a variant of the residual architecture with Shake-Shake regularization. We also use the set of optimized hyperparameters found for ResNet-18 on the CIFAR-10 dataset to train and evaluate the model on the ImageNet dataset, and show 3.43% (1.83%) improvement in the top-1 (top-5) error rate compared to the original ResNet-18 model. CURIOUS also obtains the highest accuracy for various other FFNNs that are geared toward edge devices and IoT sensors. In addition, we use CURIOUS to search for deep RNN architectures for the SICK dataset for sentence similarity evaluation. It achieves a mean-squared error of only 0.2060, improving upon the base network performance, without the need to stack multiple long short-term memories. We also use CURIOUS to search for a better NN classifier for the sentiment analysis task on the Stanford sentiment treebank dataset using a pretrained BERT model and again demonstrate improvements in performance.
引用
收藏
页码:4975 / 4990
页数:16
相关论文
共 50 条
  • [31] RENAS: Reinforced Evolutionary Neural Architecture Search
    Chen, Yukang
    Meng, Gaofeng
    Zhang, Qian
    Xiang, Shiming
    Huang, Chang
    Mu, Lisen
    Wang, Xinggang
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 4782 - 4791
  • [32] Evolutionary Neural Architecture Search for Image Restoration
    van Wyk, Gerard Jacques
    Bosman, Anna Sergeevna
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [33] Evolutionary Neural Architecture Search for Transferable Networks
    Zhou, Xun
    Liu, Songbai
    Qin, A. K.
    Tan, Kay Chen
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024,
  • [34] Guest Editorial Evolutionary Neural Architecture Search
    Sun, Yanan
    Xue, Bing
    Zhang, Mengjie
    Yen, Gary G.
    IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2024, 28 (03) : 566 - 569
  • [35] A Survey of Advances in Evolutionary Neural Architecture Search
    Zhou, Xun
    Qin, A. K.
    Sun, Yanan
    Tan, Kay Chen
    2021 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC 2021), 2021, : 950 - 957
  • [36] Evolutionary Neural Architecture Search for Traffic Forecasting
    Klosa, Daniel
    Bueskens, Christof
    2022 21ST IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS, ICMLA, 2022, : 1230 - 1237
  • [37] Search-Efficient NAS: Neural Architecture Search for Classification
    Rana, Amrita
    Kim, Kyung Ki
    2022 19TH INTERNATIONAL SOC DESIGN CONFERENCE (ISOCC), 2022, : 261 - 262
  • [38] Search-Efficient NAS: Neural Architecture Search for Classification
    Rana, Amrita
    Kim, Kyung Ki
    2022 IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS, ICDMW, 2022, : 261 - 262
  • [39] Performance Predict Method Based on Neural Architecture Search
    Zhou, Meili
    Bai, Zongwen
    Yi, Tingting
    Chen, Xiaohuan
    Wei, Wei
    JOURNAL OF INTERNET TECHNOLOGY, 2020, 21 (02): : 385 - 392
  • [40] NPENAS: Neural Predictor Guided Evolution for Neural Architecture Search
    Wei, Chen
    Niu, Chuang
    Tang, Yiping
    Wang, Yue
    Hu, Haihong
    Liang, Jimin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (11) : 8441 - 8455