An Approach of Binary Neural Network Energy-Efficient Implementation

被引:5
|
作者
Gao, Jiabao [1 ]
Liu, Qingliang [1 ]
Lai, Jinmei [1 ]
机构
[1] Fudan Univ, Sch Microelect, State Key ASIC & Syst, Shanghai 201203, Peoples R China
基金
中国国家自然科学基金;
关键词
binary neural network (BNN); kernel inclusion similarity; inclusion pruning strategy; FPGA;
D O I
10.3390/electronics10151830
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Binarized neural networks (BNNs), which have 1-bit weights and activations, are well suited for FPGA accelerators as their dominant computations are bitwise arithmetic, and the reduction in memory requirements means that all the network parameters can be stored in internal memory. However, the energy efficiency of these accelerators is still restricted by the abundant redundancies in BNNs. This hinders their deployment for applications in smart sensors and tiny devices because these scenarios have tight constraints with respect to energy consumption. To overcome this problem, we propose an approach to implement BNN inference while offering excellent energy efficiency for the accelerators by means of pruning the massive redundant operations while maintaining the original accuracy of the networks. Firstly, inspired by the observation that the convolution processes of two related kernels contain many repeated computations, we first build one formula to clarify the reusing relationships between their convolutional outputs and remove the unnecessary operations. Furthermore, by generalizing this reusing relationship to one tile of kernels in one neuron, we adopt an inclusion pruning strategy to further skip the superfluous evaluations of the neurons whose real output values can be determined early. Finally, we evaluate our system on the Zynq 7000 XC7Z100 FPGA platform. Our design can prune 51 percent of the operations without any accuracy loss. Meanwhile, the energy efficiency of our system is as high as 6.55 x 105 Img/kJ, which is 118x better than the best accelerator based on an NVDIA Tesla-V100 GPU and 3.6x higher than the state-of-the-art FPGA implementations for BNNs.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] An Energy-Efficient Systolic Pipeline Architecture for Binary Convolutional Neural Network
    Liu, Baicheng
    Chen, Song
    Kang, Yi
    Wu, Feng
    [J]. 2019 IEEE 13TH INTERNATIONAL CONFERENCE ON ASIC (ASICON), 2019,
  • [2] An Energy-Efficient FPGA-based Convolutional Neural Network Implementation
    Irmak, Hasan
    Alachiotis, Nikolaos
    Ziener, Daniel
    [J]. 29TH IEEE CONFERENCE ON SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS (SIU 2021), 2021,
  • [3] A Speculative Computation Approach for Energy-Efficient Deep Neural Network
    Zheng, Rui-Xuan
    Ko, Ya-Cheng
    Liu, Tsung-Te
    [J]. IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (03) : 795 - 806
  • [4] An Energy-Efficient Bayesian Neural Network Implementation Using Stochastic Computing Method
    Jia, Xiaotao
    Gu, Huiyi
    Liu, Yuhao
    Yang, Jianlei
    Wang, Xueyan
    Pan, Weitao
    Zhang, Youguang
    Cotofana, Sorin
    Zhao, Weisheng
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (09) : 12913 - 12923
  • [5] Designing Energy-Efficient Topologies for Wireless Sensor Network: Neural Approach
    Patra, Chiranjib
    Roy, Anjan Guha
    Chattopadhyay, Samiran
    Bhaumik, Parama
    [J]. INTERNATIONAL JOURNAL OF DISTRIBUTED SENSOR NETWORKS, 2010,
  • [6] Energy-efficient neural network chips approach human recognition capabilities
    Maass, Wolfgang
    [J]. PROCEEDINGS OF THE NATIONAL ACADEMY OF SCIENCES OF THE UNITED STATES OF AMERICA, 2016, 113 (41) : 11387 - 11389
  • [7] Towards energy-efficient neural network calculations
    Noskova, E. S.
    Zakharov, I. E.
    Shkandybin, Y. N.
    Rykovanov, S. G.
    [J]. COMPUTER OPTICS, 2022, 46 (01) : 160 - 166
  • [8] Processing-in-Memory for Energy-efficient Neural Network Training: A Heterogeneous Approach
    Liu, Jiawen
    Zhao, Hengyu
    Ogleari, Matheus Almeida
    Li, Dong
    Zhao, Jishen
    [J]. 2018 51ST ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE (MICRO), 2018, : 655 - 668
  • [9] SOT-MRAM-Based Design for Energy-Efficient and Reliable Binary Neural Network Acceleration
    Shaban, Ahmed
    Gothalyan, Shreshtha
    Hou, Tuo-Hung
    Suri, Manan
    [J]. IEEE TRANSACTIONS ON ELECTRON DEVICES, 2024, 71 (09) : 5367 - 5374
  • [10] An Energy-Efficient Time Domain Based Compute In-Memory Architecture for Binary Neural Network
    Chakraborty, Subhradip
    Kushwaha, Dinesh
    Goel, Abhishek
    Singla, Anmol
    Bulusu, Anand
    Dasgupta, Sudeb
    [J]. 2024 25TH INTERNATIONAL SYMPOSIUM ON QUALITY ELECTRONIC DESIGN, ISQED 2024, 2024,