NASA plus : Neural Architecture Search and Acceleration for Multiplication-Reduced Hybrid Networks

被引:3
|
作者
Shi, Huihong [1 ,2 ]
You, Haoran [3 ]
Wang, Zhongfeng [4 ]
Lin, Yingyan [3 ]
机构
[1] Georgia Inst Technol, Atlanta, GA 30332 USA
[2] Nanjing Univ, Sch Elect Sci & Engn, Nanjing 210093, Peoples R China
[3] Georgia Inst Technol, Sch Comp Sci, Atlanta, GA 30332 USA
[4] Nanjing Univ, Sch Elect Sci & Engn, Nanjing 210093, Peoples R China
关键词
Multiplication-reduced hybrid networks; neural architecture search; chunk-based accelerator; reconfigurable PE; algorithm-hardware co-design;
D O I
10.1109/TCSI.2023.3256700
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Multiplication is arguably the most computation-intensive operation in modern deep neural networks (DNNs), limiting their extensive deployment on resource-constrained devices. Thereby, pioneering works have handcrafted multiplication-free DNNs, which are hardware-efficient but generally inferior to their multiplication-based counterparts in task accuracy, calling for multiplication-reduced hybrid DNNs to marry the best of both worlds. To this end, we propose a Neural Architecture Search and Acceleration (NASA) framework for the above hybrid models, dubbed NASA+, to boost both task accuracy and hardware efficiency. Specifically, NASA+ augments the state-of-the-art (SOTA) search space with multiplication-free operators to construct hybrid ones, and then adopts a novel progressive pretraining strategy to enable the effective search. Furthermore, NASA+ develops a chunk-based accelerator with novel reconfigurable processing elements to better support searched hybrid models, and integrates an auto-mapper to search for optimal dataflows. Experimental results and ablation studies consistently validate the effectiveness of our NASA+ algorithm-hardware co-design framework, e.g., we can achieve up to 65.1% lower energy-delay-product with comparable accuracy over the SOTA multiplication-based system on CIFAR100. Codes are available at https://github.com/GATECH-EIC/NASA.
引用
收藏
页码:2523 / 2536
页数:14
相关论文
共 50 条
  • [11] Evaluating Acceleration Techniques for Genetic Neural Architecture Search
    Dervisi, Foteini
    Kyriakides, George
    Margaritis, Konstantinos
    ENGINEERING APPLICATIONS OF NEURAL NETWORKS, EAAAI/EANN 2022, 2022, 1600 : 3 - 14
  • [12] The effect of reduced training in neural architecture search
    George Kyriakides
    Konstantinos Margaritis
    Neural Computing and Applications, 2020, 32 : 17321 - 17332
  • [13] The effect of reduced training in neural architecture search
    Kyriakides, George
    Margaritis, Konstantinos
    NEURAL COMPUTING & APPLICATIONS, 2020, 32 (23): : 17321 - 17332
  • [14] Neural Architecture Search for Convolutional Neural Networks with Attention
    Nakai, Kohei
    Matsubara, Takashi
    Uehara, Kuniaki
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2021, E104D (02) : 312 - 321
  • [15] Neural Architecture Search of SPD Manifold Networks
    Sukthanker, Rhea Sanjay
    Huang, Zhiwu
    Kumar, Suryansh
    Endsjo, Erik Goron
    Wu, Yan
    Van Gool, Luc
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 3002 - 3009
  • [16] Evolutionary Neural Architecture Search for Transferable Networks
    Zhou, Xun
    Liu, Songbai
    Qin, A. K.
    Tan, Kay Chen
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024,
  • [17] Genetic Architecture Search for Binarized Neural Networks
    Chang, Yangyang
    Sobelman, Gerald E.
    Zhou, Xiaofang
    2019 IEEE 13TH INTERNATIONAL CONFERENCE ON ASIC (ASICON), 2019,
  • [18] Efficient Architecture Search for Deep Neural Networks
    Gottapu, Ram Deepak
    Dagli, Cihan H.
    COMPLEX ADAPTIVE SYSTEMS, 2020, 168 : 19 - 25
  • [19] Neural Architecture Search for Low-Precision Neural Networks
    Wu, Binyi
    Waschneck, Bernd
    Mayr, Christian
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2022, PT IV, 2022, 13532 : 743 - 755
  • [20] Adversarially Robust Neural Architecture Search for Graph Neural Networks
    Xie, Beini
    Chang, Heng
    Zhang, Ziwei
    Wang, Xin
    Wang, Daxin
    Zhang, Zhiqiang
    Ying, Rex
    Zhu, Wenwu
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 8143 - 8152