NASA plus : Neural Architecture Search and Acceleration for Multiplication-Reduced Hybrid Networks

被引:3
|
作者
Shi, Huihong [1 ,2 ]
You, Haoran [3 ]
Wang, Zhongfeng [4 ]
Lin, Yingyan [3 ]
机构
[1] Georgia Inst Technol, Atlanta, GA 30332 USA
[2] Nanjing Univ, Sch Elect Sci & Engn, Nanjing 210093, Peoples R China
[3] Georgia Inst Technol, Sch Comp Sci, Atlanta, GA 30332 USA
[4] Nanjing Univ, Sch Elect Sci & Engn, Nanjing 210093, Peoples R China
关键词
Multiplication-reduced hybrid networks; neural architecture search; chunk-based accelerator; reconfigurable PE; algorithm-hardware co-design;
D O I
10.1109/TCSI.2023.3256700
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Multiplication is arguably the most computation-intensive operation in modern deep neural networks (DNNs), limiting their extensive deployment on resource-constrained devices. Thereby, pioneering works have handcrafted multiplication-free DNNs, which are hardware-efficient but generally inferior to their multiplication-based counterparts in task accuracy, calling for multiplication-reduced hybrid DNNs to marry the best of both worlds. To this end, we propose a Neural Architecture Search and Acceleration (NASA) framework for the above hybrid models, dubbed NASA+, to boost both task accuracy and hardware efficiency. Specifically, NASA+ augments the state-of-the-art (SOTA) search space with multiplication-free operators to construct hybrid ones, and then adopts a novel progressive pretraining strategy to enable the effective search. Furthermore, NASA+ develops a chunk-based accelerator with novel reconfigurable processing elements to better support searched hybrid models, and integrates an auto-mapper to search for optimal dataflows. Experimental results and ablation studies consistently validate the effectiveness of our NASA+ algorithm-hardware co-design framework, e.g., we can achieve up to 65.1% lower energy-delay-product with comparable accuracy over the SOTA multiplication-based system on CIFAR100. Codes are available at https://github.com/GATECH-EIC/NASA.
引用
收藏
页码:2523 / 2536
页数:14
相关论文
共 50 条
  • [41] Federated Neural Architecture Search with Hierarchical Progressive Acceleration for Medical Image Segmentation
    Wu, Yu
    Fan, Hailong
    Ying, Weiqin
    Zhou, Zekun
    Zheng, Qiaoqiao
    Zhang, Jiajian
    ADVANCES IN SWARM INTELLIGENCE, PT II, ICSI 2024, 2024, 14789 : 112 - 123
  • [42] NASGuard: A Novel Accelerator Architecture for Robust Neural Architecture Search (NAS) Networks
    Wang, Xingbin
    Zhao, Boyan
    Hou, Rui
    Awad, Amro
    Tian, Zhihong
    Meng, Dan
    2021 ACM/IEEE 48TH ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA 2021), 2021, : 776 - 789
  • [43] Search-Free Inference Acceleration for Sparse Convolutional Neural Networks
    Liu, Bosheng
    Chen, Xiaoming
    Han, Yinhe
    Wu, Jigang
    Chang, Liang
    Liu, Peng
    Xu, Haobo
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (07) : 2156 - 2169
  • [44] Hybrid Accelerator with MapReduce Architecture for Convolutional Neural Networks
    Mihaita, David
    Stefan, Gheorghe M.
    ROMANIAN JOURNAL OF INFORMATION SCIENCE AND TECHNOLOGY, 2017, 20 (03): : 186 - 197
  • [45] Hybrid neural networks: An evolutionary approach with local search
    Iyoda, EM
    Von Zuben, FJ
    INTEGRATED COMPUTER-AIDED ENGINEERING, 2002, 9 (01) : 57 - 72
  • [46] Partially-Connected Neural Architecture Search for Reduced Computational Redundancy
    Xu, Yuhui
    Xie, Lingxi
    Dai, Wenrui
    Zhang, Xiaopeng
    Chen, Xin
    Qi, Guo-Jun
    Xiong, Hongkai
    Tian, Qi
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (09) : 2953 - 2970
  • [47] Designing Real-Time Neural Networks by Efficient Neural Architecture Search
    Bo, Zitong
    Li, Yilin
    Qiao, Ying
    Leng, Chang
    Wang, Hongan
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT IV, ICIC 2024, 2024, 14865 : 62 - 73
  • [48] Multi-objective Evolutionary Neural Architecture Search for Recurrent Neural Networks
    Booysen, Reinhard
    Bosman, Anna Sergeevna
    NEURAL PROCESSING LETTERS, 2024, 56 (04)
  • [49] Universal Binary Neural Networks Design by Improved Differentiable Neural Architecture Search
    Tan, Menghao
    Gao, Weifeng
    Li, Hong
    Xie, Jin
    Gong, Maoguo
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (10) : 9153 - 9165
  • [50] Lightweight Neural Architecture Search for Temporal Convolutional Networks at the Edge
    Risso, Matteo
    Burrello, Alessio
    Conti, Francesco
    Lamberti, Lorenzo
    Chen, Yukai
    Benini, Luca
    Macii, Enrico
    Poncino, Massimo
    Pagliari, Daniele Jahier
    IEEE TRANSACTIONS ON COMPUTERS, 2023, 72 (03) : 744 - 758