Towards High Performance and Accurate BNN Inference on FPGA with Structured Fine-grained Pruning

被引:1
|
作者
Fu, Keqi [1 ]
Qi, Zhi [1 ]
Cai, Jiaxuan [1 ]
Shi, Xulong [1 ]
机构
[1] Southeast Univ, Elect Engn Dept, Natl ASIC Syst Engn Res Ctr, Nanjing, Peoples R China
基金
美国国家科学基金会;
关键词
binary neural networks; structured pruning; accelerator architecture; FPGA;
D O I
10.1145/3508352.3549368
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
As the extreme case of quantization networks, Binary Neural Networks (BNNs) have received tremendous attention due to many hardware-friendly properties in terms of storage and computation. To reach the limit of compact models, we attempt to combine binarization with pruning techniques, further exploring the redundancy of BNNs. However, coarse-grained pruning methods may cause server accuracy drops, while traditional fine-grained ones induce irregular sparsity hard to be utilized by hardware. In this paper, we propose two advanced fine-grained BNN pruning modules, i.e., structured channel-wise kernel pruning and dynamic spatial pruning, from a joint perspective of algorithm and hardware. The pruned BNN models are trained from scratch and present not only a higher precision but also a high degree of parallelism. Then, we develop an accelerator architecture that can effectively exploit the sparsity caused by our algorithm. Finally, we implement the pruned BNN models on an embedded FPGA (Ultra96v2). The results show that our software and hardware codesign achieves 5.4x inference-speedup than the baseline BNN, with higher resource and energy efficiency compared with prior FPGA implemented BNN works.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] O3BNN: An Out-Of-Order Architecture for High-Performance Binarized Neural Network Inference with Fine-Grained Pruning
    Geng, Tong
    Wang, Tianqi
    Wu, Chunshu
    Yang, Chen
    Wu, Wei
    Li, Ang
    Herbordt, Martin C.
    INTERNATIONAL CONFERENCE ON SUPERCOMPUTING (ICS 2019), 2019, : 461 - 472
  • [2] An FPGA Overlay for CNN Inference with Fine-grained Flexible Parallelism
    Choudhury, Ziaul
    Shrivastava, Shashwat
    Ramapantulu, Lavanya
    Purini, Suresh
    ACM TRANSACTIONS ON ARCHITECTURE AND CODE OPTIMIZATION, 2022, 19 (03)
  • [3] High-performance fine-grained pipelined LMS algorithm in virtex FPGA
    Ting, LK
    Woods, RF
    Cowan, CFN
    Cork, P
    Sprigings, C
    ADVANCED SIGNAL PROCESSING ALGORITHMS, ARCHITECTURES, AND IMPLEMENTATIONS X, 2000, 4116 : 288 - 299
  • [4] Accurate and Scalable Gaussian Processes for Fine-Grained Air Quality Inference
    Patel, Zeel B.
    Purohit, Palak
    Patel, Harsh M.
    Sahni, Shivam
    Batra, Nipun
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 12080 - 12088
  • [5] Comparing Fine-Grained Performance on the Ambric MPPA against an FPGA
    Hutchings, Brad
    Nelson, Brent
    West, Stephen
    Curtis, Reed
    FPL: 2009 INTERNATIONAL CONFERENCE ON FIELD PROGRAMMABLE LOGIC AND APPLICATIONS, 2009, : 174 - 179
  • [6] Modular exponentiation on fine-grained FPGA
    Tiountchik, A
    Trichina, E
    13TH SYMPOSIUM ON INTEGRATED CIRCUITS AND SYSTEMS DESIGN, PROCEEDINGS, 2000, : 139 - 143
  • [7] Modular exponentiation on fine-grained FPGA
    Tiountchik, A
    Trichina, E
    TOPICS IN CRYPTOLOGY - CT-RAS 2001, PROCEEDINGS, 2001, 2020 : 223 - 234
  • [8] Leveraging Fine-grained Structured Sparsity for CNN Inference on Systolic Array Architectures
    Liu, Linqiao
    Brown, Stephen
    2021 31ST INTERNATIONAL CONFERENCE ON FIELD-PROGRAMMABLE LOGIC AND APPLICATIONS (FPL 2021), 2021, : 301 - 305
  • [9] Fine-Grained Urban Flow Inference
    Ouyang, Kun
    Liang, Yuxuan
    Liu, Ye
    Tong, Zekun
    Ruan, Sijie
    Zheng, Yu
    Rosenblum, David S.
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2022, 34 (06) : 2755 - 2770
  • [10] Fine-grained analysis of the transformer model for efficient pruning
    Ben Letaifa, Leila
    Rouas, Jean-Luc
    2022 21ST IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS, ICMLA, 2022, : 897 - 902