Efficient Hardware Acceleration for Approximate Inference of Bitwise Deep Neural Networks

被引:0
|
作者
Vogel, Sebastian [1 ]
Guntoro, Andre [1 ]
Ascheid, Gerd [2 ]
机构
[1] Robert Bosch GmbH, Corp Res, Renningen, Germany
[2] Rhein Westfal TH Aachen, Inst Commun Technol & Embedded Syst, Aachen, Germany
关键词
D O I
暂无
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In recent years, Deep Neural Networks (DNNs) have been of special interest in the area of image processing and scene perception. Albeit being effective and accurate, DNNs demand challenging computational resources. Fortunately, dedicated low bitwidth accelerators enable efficient, real-time inference of DNNs. We present an approximate evaluation method and a specialized multiplierless accelerator for the recently proposed bitwise DNNs. Our approximate evaluation method is based on the speculative recomputation of selective parts of a bitwise neural network. The selection is based on the intermediate results of a previous input evaluation. In context with limited energy budgets, our method and accelerator enable a fast, power efficient, first decision. If necessary, a reliable and accurate output is available after reevaluating the input data multiple times in an approximate manner. Our experiments on the GTSRB and CIFAR-10 dataset show that this approach results in no loss of classification performance in comparison with floating-point evaluation. Our work contributes to efficient inference of neural networks on power-constrained embedded devices.
引用
收藏
页数:6
相关论文
共 50 条
  • [1] Efficient Hardware Optimization Strategies for Deep Neural Networks Acceleration Chip
    Zhang Meng
    Zhang Jingwei
    Li Guoqing
    Wu Ruixia
    Zeng Xiaoyang
    [J]. JOURNAL OF ELECTRONICS & INFORMATION TECHNOLOGY, 2021, 43 (06) : 1510 - 1517
  • [2] Efficient Hardware Acceleration of Convolutional Neural Networks
    Kala, S.
    Jose, Babita R.
    Mathew, Jimson
    Nalesh, S.
    [J]. 32ND IEEE INTERNATIONAL SYSTEM ON CHIP CONFERENCE (IEEE SOCC 2019), 2019, : 191 - 192
  • [3] Quantized Deep Neural Networks for Energy Efficient Hardware-based Inference
    Ding, Ruizhou
    Liu, Zeye
    Blanton, R. D.
    Marculescu, Diana
    [J]. 2018 23RD ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2018, : 1 - 8
  • [4] A Survey on Efficient Convolutional Neural Networks and Hardware Acceleration
    Ghimire, Deepak
    Kil, Dayoung
    Kim, Seong-heum
    [J]. ELECTRONICS, 2022, 11 (06)
  • [5] DSP-Efficient Hardware Acceleration of Convolutional Neural Network Inference on FPGAs
    Wang, Dong
    Xu, Ke
    Guo, Jingning
    Ghiasi, Soheil
    [J]. IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2020, 39 (12) : 4867 - 4880
  • [6] Efficient Softmax Hardware Architecture for Deep Neural Networks
    Du, Gaoming
    Tian, Chao
    Li, Zhenmin
    Zhang, Duoli
    Yin, Yongsheng
    Ouyang, Yiming
    [J]. GLSVLSI '19 - PROCEEDINGS OF THE 2019 ON GREAT LAKES SYMPOSIUM ON VLSI, 2019, : 75 - 80
  • [7] COMPRESSING DEEP NEURAL NETWORKS FOR EFFICIENT VISUAL INFERENCE
    Ge, Shiming
    Luo, Zhao
    Zhao, Shengwei
    Jin, Xin
    Zhang, Xiao-Yu
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2017, : 667 - 672
  • [8] Layerwise Approximate Inference for Bayesian Uncertainty Estimates on Deep Neural Networks
    Zhang, Ni
    Chen, Xiaoyi
    Quan, Li
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [9] Efficient Hardware Acceleration of Sparsely Active Convolutional Spiking Neural Networks
    Sommer, Jan
    Ozkan, M. Akif
    Keszocze, Oliver
    Teich, Juergen
    [J]. IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (11) : 3767 - 3778
  • [10] Binarized Convolutional Neural Networks with Separable Filters for Efficient Hardware Acceleration
    Lin, Jeng-Hau
    Xing, Tianwei
    Zhao, Ritchie
    Zhang, Zhiru
    Srivastava, Mani
    Tu, Zhuowen
    Gupta, Rajesh K.
    [J]. 2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2017, : 344 - 352