CURIOUS: Efficient Neural Architecture Search Based on a Performance Predictor and Evolutionary Search

被引:4
|
作者
Hassantabar, Shayan [1 ]
Dai, Xiaoliang [2 ]
Jha, Niraj K. [1 ]
机构
[1] Princeton Univ, Dept Elect & Comp Engn, Princeton, NJ 08544 USA
[2] Facebook Mobile Vis, Campton, CA USA
关键词
Compression; convolutional neural network (CNN); deep learning; dimensionality reduction; long short-term memory (LSTM); neural architecture search (NAS); transformer;
D O I
10.1109/TCAD.2022.3148202
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Neural networks (NNs) have been successfully deployed in various applications of artificial intelligence. However, architectural design of NNs is still a challenging problem. This is due to the need to navigate a search space based on a large number of hyperparameters. This forces the search space of possible architectures to grow exponentially. Using a trial-and-error design approach is very time consuming and leads to suboptimal architectures. In addition, approaches, such as neural architecture search based on reinforcement learning and differentiable gradient-based architecture search, often incur huge computational costs or significant memory requirements. To address these challenges, we propose the CURIOUS NN synthesis methodology. It uses a performance predictor to efficiently navigate the architectural search space with an evolutionary search process. The predictor is built using quasi Monte-Carlo sampling, boosted decision tree regression, and an intelligent iterative sampling method. It is designed to be sample efficient. CURIOUS starts from a base architecture and explores the architectural search space to obtain a variant of the base architecture with the highest performance. This search framework is general and covers all important NN architecture types, e.g., feedforward NNs (FFNNs), convolutional NNs (CNNs), recurrent NNs (RNNs), and transformers. We evaluate the performance of CURIOUS on various datasets and base architectures. Through these experiments, we demonstrate significant performance improvements over the baseline architectures. For theMNIST dataset, our CNN architecture achieves an error rate of 0.66%, with 8.6x fewer parameters compared to the LeNet-5 baseline. For the CIFAR-10 dataset, we use the ResNet architectures and residual networks with Shake-Shake regularization as the baselines. Our synthesized ResNet-18 has a 2.52% accuracy improvement over the original ResNet-18, 1.74% over ResNet-101, and 0.16% over ResNet-1001, while requiring comparable number of parameters and floating-point operations to the original ResNet-18. This result shows that instead of just increasing the number of layers to increase accuracy, an alternative is to use a better NN architecture with a small number of layers. In addition, CURIOUS achieves an error rate of just 2.69% with a variant of the residual architecture with Shake-Shake regularization. We also use the set of optimized hyperparameters found for ResNet-18 on the CIFAR-10 dataset to train and evaluate the model on the ImageNet dataset, and show 3.43% (1.83%) improvement in the top-1 (top-5) error rate compared to the original ResNet-18 model. CURIOUS also obtains the highest accuracy for various other FFNNs that are geared toward edge devices and IoT sensors. In addition, we use CURIOUS to search for deep RNN architectures for the SICK dataset for sentence similarity evaluation. It achieves a mean-squared error of only 0.2060, improving upon the base network performance, without the need to stack multiple long short-term memories. We also use CURIOUS to search for a better NN classifier for the sentiment analysis task on the Stanford sentiment treebank dataset using a pretrained BERT model and again demonstrate improvements in performance.
引用
收藏
页码:4975 / 4990
页数:16
相关论文
共 50 条
  • [21] Effective, Efficient and Robust Neural Architecture Search Effective, Efficient and Robust Neural Architecture Search
    Yue, Zhixiong
    Lin, Baijiong
    Zhang, Yu
    Liang, Christy
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [22] Hybrid Architecture-Based Evolutionary Robust Neural Architecture Search
    Yang, Shangshang
    Sun, Xiangkun
    Xu, Ke
    Liu, Yuanchao
    Tian, Ye
    Zhang, Xingyi
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, 8 (04): : 2919 - 2934
  • [23] Quantum Architecture Search with Neural Predictor Based on Graph Measures
    He, Zhimin
    Li, Zhengjiang
    Deng, Maijie
    Zheng, Shenggen
    Situ, Haozhen
    Li, Lvzhou
    ADVANCED QUANTUM TECHNOLOGIES, 2024, 7 (11)
  • [24] EGNAS: Efficient Graph Neural Architecture Search Through Evolutionary Algorithm
    Jwa, Younkyung
    Ahn, Chang Wook
    Kim, Man-Je
    MATHEMATICS, 2024, 12 (23)
  • [25] Neural Architecture Search Based on Evolutionary Algorithms with Fitness Approximation
    Pan, Chao
    Yao, Xin
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [26] Automated design of CNN architecture based on efficient evolutionary search
    Xie, Yirong
    Chen, Hong
    Ma, Yongjie
    Xu, Yang
    NEUROCOMPUTING, 2022, 491 : 160 - 171
  • [27] Novelty Driven Evolutionary Neural Architecture Search
    Sinha, Nilotpal
    Chen, Kuan-Wen
    PROCEEDINGS OF THE 2022 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION, GECCO 2022, 2022, : 671 - 674
  • [28] Benchmarking Analysis of Evolutionary Neural Architecture Search
    Lv, Zeqiong
    Qian, Chao
    Sun, Yanan
    IEEE TRANSACTIONS ON EVOLUTIONARY COMPUTATION, 2024, 28 (06) : 1659 - 1673
  • [29] Efficient Evolution for Neural Architecture Search
    Chen, Zihao
    Li, Bin
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [30] Efficient Global Neural Architecture Search
    Shahid Siddiqui
    Christos Kyrkou
    Theocharis Theocharides
    SN Computer Science, 6 (3)