NASA plus : Neural Architecture Search and Acceleration for Multiplication-Reduced Hybrid Networks

被引:3
|
作者
Shi, Huihong [1 ,2 ]
You, Haoran [3 ]
Wang, Zhongfeng [4 ]
Lin, Yingyan [3 ]
机构
[1] Georgia Inst Technol, Atlanta, GA 30332 USA
[2] Nanjing Univ, Sch Elect Sci & Engn, Nanjing 210093, Peoples R China
[3] Georgia Inst Technol, Sch Comp Sci, Atlanta, GA 30332 USA
[4] Nanjing Univ, Sch Elect Sci & Engn, Nanjing 210093, Peoples R China
关键词
Multiplication-reduced hybrid networks; neural architecture search; chunk-based accelerator; reconfigurable PE; algorithm-hardware co-design;
D O I
10.1109/TCSI.2023.3256700
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Multiplication is arguably the most computation-intensive operation in modern deep neural networks (DNNs), limiting their extensive deployment on resource-constrained devices. Thereby, pioneering works have handcrafted multiplication-free DNNs, which are hardware-efficient but generally inferior to their multiplication-based counterparts in task accuracy, calling for multiplication-reduced hybrid DNNs to marry the best of both worlds. To this end, we propose a Neural Architecture Search and Acceleration (NASA) framework for the above hybrid models, dubbed NASA+, to boost both task accuracy and hardware efficiency. Specifically, NASA+ augments the state-of-the-art (SOTA) search space with multiplication-free operators to construct hybrid ones, and then adopts a novel progressive pretraining strategy to enable the effective search. Furthermore, NASA+ develops a chunk-based accelerator with novel reconfigurable processing elements to better support searched hybrid models, and integrates an auto-mapper to search for optimal dataflows. Experimental results and ablation studies consistently validate the effectiveness of our NASA+ algorithm-hardware co-design framework, e.g., we can achieve up to 65.1% lower energy-delay-product with comparable accuracy over the SOTA multiplication-based system on CIFAR100. Codes are available at https://github.com/GATECH-EIC/NASA.
引用
收藏
页码:2523 / 2536
页数:14
相关论文
共 50 条
  • [31] Evolving graph convolutional networks for neural architecture search
    Kyriakides, George
    Margaritis, Konstantinos
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (02): : 899 - 909
  • [32] DASS: Differentiable Architecture Search for Sparse Neural Networks
    Mousavi, Hamid
    Loni, Mohammad
    Alibeigi, Mina
    Daneshtalab, Masoud
    ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2023, 22 (05)
  • [33] Evolving graph convolutional networks for neural architecture search
    George Kyriakides
    Konstantinos Margaritis
    Neural Computing and Applications, 2022, 34 : 899 - 909
  • [34] GraphPAS: Parallel Architecture Search for Graph Neural Networks
    Chen, Jiamin
    Gao, Jianliang
    Chen, Yibo
    Oloulade, Moctard Babatounde
    Lyu, Tengfei
    Li, Zhao
    SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, : 2182 - 2186
  • [35] Auto-GNN: Neural architecture search of graph neural networks
    Zhou, Kaixiong
    Huang, Xiao
    Song, Qingquan
    Chen, Rui
    Hu, Xia
    FRONTIERS IN BIG DATA, 2022, 5
  • [36] EDANAS: Adaptive Neural Architecture Search for Early Exit Neural Networks
    Gambella, Matteo
    Roveri, Manuel
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [37] A surrogate evolutionary neural architecture search algorithm for graph neural networks
    Liu, Yang
    Liu, Jing
    APPLIED SOFT COMPUTING, 2023, 144
  • [38] NAS-BNN: Neural Architecture Search for Binary Neural Networks
    Lin, Zhihao
    Wang, Yongtao
    Zhang, Jinhe
    Chu, Xiaojie
    Ling, Haibin
    PATTERN RECOGNITION, 2025, 159
  • [39] Hybrid classical–quantum neural networks enhanced by quantum architecture search for coronary artery stenosis detection
    Li, Shaochun
    Cui, Junzhi
    Ren, Jingli
    Neurocomputing, 2025, 618
  • [40] Training Acceleration for Deep Neural Networks: A Hybrid Parallelization Strategy
    Zeng, Zihao
    Liu, Chubo
    Tang, Zhuo
    Chang, Wanli
    Li, Kenli
    2021 58TH ACM/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2021, : 1165 - 1170