Accelerating one-shot neural architecture search via constructing a sparse search space

被引:0
|
作者
Huang, Hongtao [1 ]
Chang, Xiaojun [2 ]
Yao, Lina [1 ,3 ]
机构
[1] School of Computer Science and Engineering, University of New South Wales, Kensington, Sydney,NSW,2052, Australia
[2] Faculty of Engineering & Information Technology, University of Technology Sydney, Ultimo, Sydney,NSW,2007, Australia
[3] CSIRO's Data61, Eveleigh, Sydney,NSW,2015, Australia
关键词
Digital storage - Graphics processing unit;
D O I
10.1016/j.knosys.2024.112620
中图分类号
学科分类号
摘要
Neural Architecture Search (NAS) has garnered significant attention for its ability to automatically design high-quality deep neural networks (DNNs) tailored to various hardware platforms. The major challenge for NAS is the time-consuming network estimation process required to select optimal networks from a large pool of candidates. Rather than training each candidate from scratch, recent one-shot NAS methods accelerate the estimation process by only training a supernet and sampling sub-networks from it, inheriting partial network architectures and weights. Despite significant acceleration, the supernet training with a large search space (i.e., the number of candidate sub-networks) still requires thousands of GPU hours to support high-quality sub-network sampling. In this work, we propose SparseNAS, an approach for one-shot NAS acceleration by reducing the redundancy of the search space. We observe that many sub-networks in the space are underperforming, with significant performance disparity to high-performance sub-networks. Crucially, this disparity can be observed early in the beginning of the supernet training. Therefore, we train an early predictor to learn this disparity and filter out high-quality networks in advance. Then, the supernet training will be conducted in this space sub-space. Compared to the state-of-the-art one-shot NAS, our SparseNAS reports a 3.1× training speedup with comparable network performance on the ImageNet dataset. Compared to the state-of-the-art acceleration method, SparseNAS reports a maximum of 1.5% higher Top-1 accuracy and 28% training cost reduction with a 7× bigger search space. Extensive experiment results demonstrated that SparseNAS achieves better trade-offs between efficiency and performance than state-of-the-art one-shot NAS. © 2024 The Authors
引用
收藏
相关论文
共 50 条
  • [1] One-shot Graph Neural Architecture Search with Dynamic Search Space
    Li, Yanxi
    Wen, Zean
    Wang, Yunhe
    Xu, Chang
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 8510 - 8517
  • [2] Progressive Automatic Design of Search Space for One-Shot Neural Architecture Search
    Xia, Xin
    Xiao, Xuefeng
    Wang, Xing
    Zheng, Min
    [J]. 2022 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2022), 2022, : 3525 - 3534
  • [3] One-Shot Neural Architecture Search via Novelty Driven Sampling
    Zhang, Miao
    Li, Huiqi
    Pan, Shirui
    Liu, Taoping
    Su, Steven
    [J]. PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 3188 - 3194
  • [4] One-Shot Neural Ensemble Architecture Search by Diversity-Guided Search Space Shrinking
    Chen, Minghao
    Fu, Jianlong
    Ling, Haibin
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 16525 - 16534
  • [5] MixPath: A Unified Approach for One-shot Neural Architecture Search
    Chu, Xiangxiang
    Lu, Shun
    Li, Xudong
    Zhang, Bo
    [J]. 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 5949 - 5958
  • [6] One-Shot Neural Architecture Search via Self-Evaluated Template Network
    Dong, Xuanyi
    Yang, Yi
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 3680 - 3689
  • [7] Understanding and Simplifying One-Shot Architecture Search
    Bender, Gabriel
    Kindermans, Pieter-Jan
    Zoph, Barret
    Vasudevan, Vijay
    Le, Quoc
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [8] Optimized one-shot neural architecture search for skin cancer classification
    Damarla, Anupama
    Doraikannan, Sumathi
    [J]. JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (06)
  • [9] LightTrack: Finding Lightweight Neural Networks for Object Tracking via One-Shot Architecture Search
    Yan, Bin
    Peng, Houwen
    Wu, Kan
    Wang, Dong
    Fu, Jianlong
    Lu, Huchuan
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 15175 - 15184
  • [10] Layer-Wisely Supervised Learning For One-Shot Neural Architecture Search
    Chen, Yifei
    Guo, Zhourui
    Yin, Qiyue
    Chen, Hao
    Huang, Kaiqi
    [J]. 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,