Accelerating one-shot neural architecture search via constructing a sparse search space

被引:0
|
作者
Huang, Hongtao [1 ]
Chang, Xiaojun [2 ]
Yao, Lina [1 ,3 ]
机构
[1] School of Computer Science and Engineering, University of New South Wales, Kensington, Sydney,NSW,2052, Australia
[2] Faculty of Engineering & Information Technology, University of Technology Sydney, Ultimo, Sydney,NSW,2007, Australia
[3] CSIRO's Data61, Eveleigh, Sydney,NSW,2015, Australia
关键词
Digital storage - Graphics processing unit;
D O I
10.1016/j.knosys.2024.112620
中图分类号
学科分类号
摘要
Neural Architecture Search (NAS) has garnered significant attention for its ability to automatically design high-quality deep neural networks (DNNs) tailored to various hardware platforms. The major challenge for NAS is the time-consuming network estimation process required to select optimal networks from a large pool of candidates. Rather than training each candidate from scratch, recent one-shot NAS methods accelerate the estimation process by only training a supernet and sampling sub-networks from it, inheriting partial network architectures and weights. Despite significant acceleration, the supernet training with a large search space (i.e., the number of candidate sub-networks) still requires thousands of GPU hours to support high-quality sub-network sampling. In this work, we propose SparseNAS, an approach for one-shot NAS acceleration by reducing the redundancy of the search space. We observe that many sub-networks in the space are underperforming, with significant performance disparity to high-performance sub-networks. Crucially, this disparity can be observed early in the beginning of the supernet training. Therefore, we train an early predictor to learn this disparity and filter out high-quality networks in advance. Then, the supernet training will be conducted in this space sub-space. Compared to the state-of-the-art one-shot NAS, our SparseNAS reports a 3.1× training speedup with comparable network performance on the ImageNet dataset. Compared to the state-of-the-art acceleration method, SparseNAS reports a maximum of 1.5% higher Top-1 accuracy and 28% training cost reduction with a 7× bigger search space. Extensive experiment results demonstrated that SparseNAS achieves better trade-offs between efficiency and performance than state-of-the-art one-shot NAS. © 2024 The Authors
引用
收藏
相关论文
共 50 条
  • [31] Neural Architecture Search as Sparse Supernet
    Wu, Yan
    Liu, Aoming
    Huang, Zhiwu
    Zhang, Siwei
    Van Gool, Luc
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 10379 - 10387
  • [32] One-Shot Informed Robotic Visual Search in the Wild
    Koreitem, Karim
    Shkurti, Florian
    Manderson, Travis
    Chang, Wei-Di
    Higuera, Juan Camilo Gamboa
    Dudek, Gregory
    [J]. 2020 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2020, : 5800 - 5807
  • [33] HyperSegNAS: Bridging One-Shot Neural Architecture Search with 3D Medical Image Segmentation using HyperNet
    Peng, Cheng
    Myronenko, Andriy
    Hatamizadeh, Ali
    Nath, Vishwesh
    Siddiquee, Md Mahfuzur Rahman
    He, Yufan
    Xu, Daguang
    Chellappa, Rama
    Yang, Dong
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 20709 - 20719
  • [34] Few-shot Neural Architecture Search
    Zhao, Yiyang
    Wang, Linnan
    Tian, Yuandong
    Fonseca, Rodrigo
    Guo, Tian
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [35] Exploring Neural Architecture Search Space via Deep Deterministic Sampling
    Mills, Keith G.
    Salameh, Mohammad
    Niu, Di
    Han, Fred X.
    Rezaei, Seyed Saeed Changiz
    Yao, Hengshuai
    Lu, Wei
    Lian, Shuo
    Jui, Shangling
    [J]. IEEE ACCESS, 2021, 9 : 110962 - 110974
  • [36] DASS: Differentiable Architecture Search for Sparse Neural Networks
    Mousavi, Hamid
    Loni, Mohammad
    Alibeigi, Mina
    Daneshtalab, Masoud
    [J]. ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2023, 22 (05)
  • [37] Neural Architecture Search for Sparse DenseNets with Dynamic Compression
    O'Neill, Damien
    Xue, Bing
    Zhang, Mengjie
    [J]. GECCO'20: PROCEEDINGS OF THE 2020 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE, 2020, : 386 - 394
  • [38] One-Shot Neural Face Reenactment via Finding Directions in GAN's Latent Space
    Bounareli, Stella
    Tzelepis, Christos
    Argyriou, Vasileios
    Patras, Ioannis
    Tzimiropoulos, Georgios
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2024, 132 (08) : 3324 - 3354
  • [39] Search Space Adaptation for Differentiable Neural Architecture Search in Image Classification
    Kim, Youngkee
    Jung, Soyi
    Choi, Minseok
    Kim, Joongheon
    [J]. 2022 THIRTEENTH INTERNATIONAL CONFERENCE ON UBIQUITOUS AND FUTURE NETWORKS (ICUFN), 2022, : 363 - 365
  • [40] Efficient evolutionary neural architecture search based on hybrid search space
    Gong, Tao
    Ma, Yongjie
    Xu, Yang
    Song, Changwei
    [J]. INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (08) : 3313 - 3326