Accelerating Sparse Deep Neural Networks on FPGAs

被引:0
|
作者
Huang, Sitao [1 ]
Pearson, Carl [1 ]
Nagi, Rakesh [1 ]
Xiong, Jinjun [2 ]
Chen, Deming [1 ]
Hwu, Wen-mei [1 ]
机构
[1] Univ Illinois, Champaign, IL 61820 USA
[2] IBM Res, Armonk, NY USA
关键词
Deep learning; Sparse DNN; Graphs; FPGA;
D O I
暂无
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks (DNNs) have been widely adopted in many domains, including computer vision, natural language processing, and medical care. Recent research reveals that sparsity in DNN parameters can be exploited to reduce inference computational complexity and improve network quality. However, sparsity also introduces irregularity and extra complexity in data processing, which make the accelerator design challenging. This work presents the design and implementation of a highly flexible sparse DNN inference accelerator on FPGA. Our proposed inference engine can be easily configured to be used in both mobile computing and high-performance computing scenarios. Evaluation shows our proposed inference engine effectively accelerates sparse DNNs and outperforms CPU solution by up to 4:7x in terms of energy efficiency.
引用
收藏
页数:7
相关论文
共 50 条
  • [1] Accelerating Deep Neural Networks Using FPGAs and ZYNQ
    Lee, Han Sung
    Jeon, Jae Wook
    [J]. 2021 IEEE REGION 10 SYMPOSIUM (TENSYMP), 2021,
  • [2] TensorFlow to Cloud FPGAs: Tradeoffs for Accelerating Deep Neural Networks
    Hadjis, Stefan
    Olukotun, Kunle
    [J]. 2019 29TH INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE LOGIC AND APPLICATIONS (FPL), 2019, : 360 - 366
  • [3] Exploring Heterogeneous Algorithms for Accelerating Deep Convolutional Neural Networks on FPGAs
    Xiao, Qincheng
    Liang, Yun
    Lu, Liqiang
    Yan, Shengen
    Tai, Yu-Wing
    [J]. PROCEEDINGS OF THE 2017 54TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2017,
  • [4] EmbRace: Accelerating Sparse Communication for Distributed Training of Deep Neural Networks
    Li, Shengwei
    Lai, Zhiquan
    Li, Dongsheng
    Zhang, Yiming
    Ye, Xiangyu
    Duan, Yabo
    [J]. 51ST INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING, ICPP 2022, 2022,
  • [5] Accelerating Training of Deep Neural Networks via Sparse Edge Processing
    Dey, Sourya
    Shao, Yinan
    Chugg, Keith M.
    Beerel, Peter A.
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2017, PT I, 2017, 10613 : 273 - 280
  • [6] Can FPGAs Beat GPUs in Accelerating Next-Generation Deep Neural Networks?
    Nurvitadhi, Eriko
    Venkatesh, Ganesh
    Sim, Jaewoong
    Marr, Debbie
    Huang, Randy
    Ong, Jason Gee Hock
    Liew, Yeong Tat
    Srivatsan, Krishnan
    Moss, Duncan
    Subhaschandra, Suchit
    Boudoukh, Guy
    [J]. FPGA'17: PROCEEDINGS OF THE 2017 ACM/SIGDA INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE GATE ARRAYS, 2017, : 5 - 14
  • [7] SyncNN: Evaluating and Accelerating Spiking Neural Networks on FPGAs
    Panchapakesan, Sathish
    Fang, Zhenman
    Li, Jian
    [J]. 2021 31ST INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE LOGIC AND APPLICATIONS (FPL 2021), 2021, : 286 - 293
  • [8] SyncNN: Evaluating and Accelerating Spiking Neural Networks on FPGAs
    Panchapakesan, Sathish
    Fang, Zhenman
    Li, Jian
    [J]. ACM TRANSACTIONS ON RECONFIGURABLE TECHNOLOGY AND SYSTEMS, 2022, 15 (04)
  • [9] Designing and Accelerating Spiking Neural Networks using OpenCL for FPGAs
    Podobas, Artur
    Matsuoka, Satoshi
    [J]. 2017 INTERNATIONAL CONFERENCE ON FIELD PROGRAMMABLE TECHNOLOGY (ICFPT), 2017, : 255 - 258
  • [10] Accelerating Distributed Inference of Sparse Deep Neural Networks via Mitigating the Straggler Effect
    Mofrad, Mohammad Hasanzadeh
    Melhem, Rami
    Ahmad, Yousuf
    Hammoud, Mohammad
    [J]. 2020 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE (HPEC), 2020,