NASA plus : Neural Architecture Search and Acceleration for Multiplication-Reduced Hybrid Networks

被引:3
|
作者
Shi, Huihong [1 ,2 ]
You, Haoran [3 ]
Wang, Zhongfeng [4 ]
Lin, Yingyan [3 ]
机构
[1] Georgia Inst Technol, Atlanta, GA 30332 USA
[2] Nanjing Univ, Sch Elect Sci & Engn, Nanjing 210093, Peoples R China
[3] Georgia Inst Technol, Sch Comp Sci, Atlanta, GA 30332 USA
[4] Nanjing Univ, Sch Elect Sci & Engn, Nanjing 210093, Peoples R China
关键词
Multiplication-reduced hybrid networks; neural architecture search; chunk-based accelerator; reconfigurable PE; algorithm-hardware co-design;
D O I
10.1109/TCSI.2023.3256700
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Multiplication is arguably the most computation-intensive operation in modern deep neural networks (DNNs), limiting their extensive deployment on resource-constrained devices. Thereby, pioneering works have handcrafted multiplication-free DNNs, which are hardware-efficient but generally inferior to their multiplication-based counterparts in task accuracy, calling for multiplication-reduced hybrid DNNs to marry the best of both worlds. To this end, we propose a Neural Architecture Search and Acceleration (NASA) framework for the above hybrid models, dubbed NASA+, to boost both task accuracy and hardware efficiency. Specifically, NASA+ augments the state-of-the-art (SOTA) search space with multiplication-free operators to construct hybrid ones, and then adopts a novel progressive pretraining strategy to enable the effective search. Furthermore, NASA+ develops a chunk-based accelerator with novel reconfigurable processing elements to better support searched hybrid models, and integrates an auto-mapper to search for optimal dataflows. Experimental results and ablation studies consistently validate the effectiveness of our NASA+ algorithm-hardware co-design framework, e.g., we can achieve up to 65.1% lower energy-delay-product with comparable accuracy over the SOTA multiplication-based system on CIFAR100. Codes are available at https://github.com/GATECH-EIC/NASA.
引用
收藏
页码:2523 / 2536
页数:14
相关论文
共 50 条
  • [1] NASA-F: FPGA-Oriented Search and Acceleration for Multiplication-Reduced Hybrid Networks
    Shi, Huihong
    Xu, Yang
    Wang, Yuefei
    Mao, Wendong
    Wang, Zhongfeng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2024, 71 (01) : 306 - 319
  • [2] NASH: Neural Architecture and Accelerator Search for Multiplication-Reduced Hybrid Models
    Xu, Yang
    Shi, Huihong
    Wang, Zhongfeng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2024, 71 (12) : 5956 - 5968
  • [3] NASA: Neural Architecture Search and Acceleration for Hardware Inspired Hybrid Networks
    Shi, Huihong
    You, Haoran
    Zhao, Yang
    Wang, Zhongfeng
    Lin, Yingyan
    2022 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2022,
  • [4] NASA: Neural Architecture Search and Acceleration for Hardware Inspired Hybrid Networks
    Shi, Huihong
    You, Haoran
    Zhao, Yang
    Wang, Zhongfeng
    Lin, Yingyan
    2022 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, ICCAD, 2022,
  • [5] GraphNAS plus plus : Distributed Architecture Search for Graph Neural Networks
    Gao, Yang
    Zhang, Peng
    Yang, Hong
    Zhou, Chuan
    Hu, Yue
    Tian, Zhihong
    Li, Zhao
    Zhou, Jingren
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (07) : 6973 - 6987
  • [6] Neural Architecture Search Applied to Hybrid Morphological Neural Networks
    Gomes Weil, Victor Alexandre
    Florindo, Joao Batista
    INTELLIGENT SYSTEMS, PT II, 2022, 13654 : 631 - 645
  • [7] HGNAS plus plus : Efficient Architecture Search for Heterogeneous Graph Neural Networks
    Gao, Yang
    Zhang, Peng
    Zhou, Chuan
    Yang, Hong
    Li, Zhao
    Hu, Yue
    Yu, Philip S. S.
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (09) : 9448 - 9461
  • [8] Fully Pipelined FPGA Acceleration of Binary Convolutional Neural Networks with Neural Architecture Search
    Ji, Mengfei
    Al-Ars, Zaid
    Chang, Yuchun
    Zhang, Baolin
    JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2024, 33 (10)
  • [9] Trends in Neural Architecture Search: Towards the Acceleration of Search
    Kim, Youngkee
    Yun, Won Joon
    Lee, Youn Kyu
    Jung, Soyi
    Kim, Joongheon
    12TH INTERNATIONAL CONFERENCE ON ICT CONVERGENCE (ICTC 2021): BEYOND THE PANDEMIC ERA WITH ICT CONVERGENCE INNOVATION, 2021, : 421 - 424
  • [10] Neural Architecture Search for Spiking Neural Networks
    Kim, Youngeun
    Li, Yuhang
    Park, Hyoungseob
    Venkatesha, Yeshwanth
    Panda, Priyadarshini
    COMPUTER VISION, ECCV 2022, PT XXIV, 2022, 13684 : 36 - 56