EAT-NAS: elastic architecture transfer for accelerating large-scale neural architecture search

被引:0
|
作者
Jiemin Fang
Yukang Chen
Xinbang Zhang
Qian Zhang
Chang Huang
Gaofeng Meng
Wenyu Liu
Xinggang Wang
机构
[1] Huazhong University of Science and Technology,Institute of Artificial Intelligence
[2] Huazhong University of Science and Technology,School of Electronic Information and Communications
[3] Horizon Robotics,National Laboratory of Pattern Recognition, Institute of Automation
[4] Chinese Academy of Sciences,undefined
来源
关键词
architecture transfer; neural architecture search; evolutionary algorithm; large-scale dataset;
D O I
暂无
中图分类号
学科分类号
摘要
Neural architecture search (NAS) methods have been proposed to relieve human experts from tedious architecture engineering. However, most current methods are constrained in small-scale search owing to the issue of huge computational resource consumption. Meanwhile, the direct application of architectures searched on small datasets to large datasets often bears no performance guarantee due to the discrepancy between different datasets. This limitation impedes the wide use of NAS on large-scale tasks. To overcome this obstacle, we propose an elastic architecture transfer mechanism for accelerating large-scale NAS (EAT-NAS). In our implementations, the architectures are first searched on a small dataset, e.g., CIFAR-10. The best one is chosen as the basic architecture. The search process on a large dataset, e.g., ImageNet, is initialized with the basic architecture as the seed. The large-scale search process is accelerated with the help of the basic architecture. We propose not only a NAS method but also a mechanism for architecture-level transfer learning. In our experiments, we obtain two final models EATNet-A and EATNet-B, which achieve competitive accuracies of 75.5% and 75.6%, respectively, on ImageNet. Both the models also surpass the models searched from scratch on ImageNet under the same settings. For the computational cost, EAT-NAS takes only fewer than 5 days using 8 TITAN X GPUs, which is significantly less than the computational consumption of the state-of-the-art large-scale NAS methods.
引用
收藏
相关论文
共 50 条
  • [1] EAT-NAS: elastic architecture transfer for accelerating large-scale neural architecture search
    Jiemin FANG
    Yukang CHEN
    Xinbang ZHANG
    Qian ZHANG
    Chang HUANG
    Gaofeng MENG
    Wenyu LIU
    Xinggang WANG
    [J]. Science China(Information Sciences), 2021, 64 (09) : 103 - 115
  • [2] EAT-NAS: elastic architecture transfer for accelerating large-scale neural architecture search
    Fang, Jiemin
    Chen, Yukang
    Zhang, Xinbang
    Zhang, Qian
    Huang, Chang
    Meng, Gaofeng
    Liu, Wenyu
    Wang, Xinggang
    [J]. SCIENCE CHINA-INFORMATION SCIENCES, 2021, 64 (09)
  • [3] Large-Scale Graph Neural Architecture Search
    Guan, Chaoyu
    Wang, Xin
    Chen, Hong
    Zhang, Ziwei
    Zhu, Wenwu
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [4] NAS-SE: Designing A Highly-Efficient In-Situ Neural Architecture Search Engine for Large-Scale Deployment
    Wan, Qiyu
    Wang, Lening
    Wang, Jing
    Song, Shuaiwen Leon
    Fu, Xin
    [J]. 56TH IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE, MICRO 2023, 2023, : 756 - 768
  • [5] NASTransfer: Analyzing Architecture Transferability in Large Scale Neural Architecture Search
    Panda, Rameswar
    Merler, Michele
    Jaiswal, Mayoore S.
    Wu, Hui
    Ramakrishnan, Kandan
    Finkler, Ulrich
    Chen, Chun-Fu Richard
    Cho, Minsik
    Feris, Rogerio
    Kung, David
    Bhattacharjee, Bishwaranjan
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 9294 - 9302
  • [6] μNAS: Constrained Neural Architecture Search for Microcontrollers
    Liberis, Edgar
    Dudziak, Lukasz
    Lane, Nicholas D.
    [J]. PROCEEDINGS OF THE 1ST WORKSHOP ON MACHINE LEARNING AND SYSTEMS (EUROMLSYS'21), 2021, : 70 - 79
  • [7] Search-Efficient NAS: Neural Architecture Search for Classification
    Rana, Amrita
    Kim, Kyung Ki
    [J]. 2022 19TH INTERNATIONAL SOC DESIGN CONFERENCE (ISOCC), 2022, : 261 - 262
  • [8] Search-Efficient NAS: Neural Architecture Search for Classification
    Rana, Amrita
    Kim, Kyung Ki
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON DATA MINING WORKSHOPS, ICDMW, 2022, : 261 - 262
  • [9] Accelerating DNN Architecture Search at Scale Using Selective Weight Transfer
    Liu, Hongyuan
    Nicolae, Bogdan
    Di, Sheng
    Cappello, Franck
    Jog, Adwait
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON CLUSTER COMPUTING (CLUSTER 2021), 2021, : 82 - 93
  • [10] NASGuard: A Novel Accelerator Architecture for Robust Neural Architecture Search (NAS) Networks
    Wang, Xingbin
    Zhao, Boyan
    Hou, Rui
    Awad, Amro
    Tian, Zhihong
    Meng, Dan
    [J]. 2021 ACM/IEEE 48TH ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA 2021), 2021, : 776 - 789