Hardware-Driven Nonlinear Activation for Stochastic Computing Based Deep Convolutional Neural Networks

被引:0
|
作者
Li, Ji [1 ]
Yuan, Zihao [1 ]
Li, Zhe [2 ]
Ding, Caiwen [2 ]
Ren, Ao [2 ]
Qiu, Qinru [2 ]
Draper, Jeffrey [1 ,3 ]
Wang, Yanzhi [2 ]
机构
[1] Univ Southern Calif, Dept Elect Engn, Los Angeles, CA 90007 USA
[2] Syracuse Univ, Coll Engn & Comp Sci, Syracuse, NY USA
[3] Univ Southern Calif, Informat Sci Inst, Marina Del Rey, CA 90292 USA
关键词
Deep Convolutional Neural Networks; Stochastic Computing; Deep Learning; Activation Function;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, Deep Convolutional Neural Networks (DCNNs) have made unprecedented progress, achieving the accuracy close to, or even better than human-level perception in various tasks. There is a timely need to map the latest software DCNNs to application-specific hardware, in order to achieve orders of magnitude improvement in performance, energy efficiency and compactness. Stochastic Computing (SC), as a low-cost alternative to the conventional binary computing paradigm, has the potential to enable massively parallel and highly scalable hardware implementation of DCNNs. One major challenge in SC based DCNNs is designing accurate nonlinear activation functions, which have a significant impact on the network-level accuracy but cannot be implemented accurately by existing SC computing blocks. In this paper, we design and optimize SC based neurons, and we propose highly accurate activation designs for the three most frequently used activation functions in software DCNNs, i.e, hyperbolic tangent, logistic, and rectified linear units. Experimental results on LeNet-5 using MNIST dataset demonstrate that compared with a binary ASIC hardware DCNN, the DCNN with the proposed SC neurons can achieve up to 61X, 151X, and 2X improvement in terms of area, power, and energy, respectively, at the cost of small precision degradation. In addition, the SC approach achieves up to 21X and 41X of the area, 41X and 72X of the power, and 198200X and 96443X of the energy, compared with CPU and GPU approaches, respectively, while the error is increased by less than 3.07%. ReLU activation is suggested for future SC based DCNNs considering its superior performance under a small bit stream length.
引用
收藏
页码:1230 / 1236
页数:7
相关论文
共 50 条
  • [21] Accelerating Deep Convolutional Neural Network base on stochastic computing
    Sadi, Mohamad Hasani
    Mahani, Ali
    INTEGRATION-THE VLSI JOURNAL, 2021, 76 : 113 - 121
  • [22] In-Memory Computing Based Hardware Accelerator Module for Deep Neural Networks
    Appukuttan, Allen
    Thomas, Emmanuel
    Nair, Harinandan R.
    Hemanth, S.
    Dhanaraj, K. J.
    Azeez, Maleeha Abdul
    2022 IEEE 19TH INDIA COUNCIL INTERNATIONAL CONFERENCE, INDICON, 2022,
  • [23] Stochastic Selection of Activation Layers for Convolutional Neural Networks
    Nanni, Loris
    Lumini, Alessandra
    Ghidoni, Stefano
    Maguolo, Gianluca
    SENSORS, 2020, 20 (06)
  • [24] Scalable Stochastic-Computing Accelerator for Convolutional Neural Networks
    Sim, Hyeonuk
    Dong Nguyen
    Lee, Jongeun
    Choi, Kiyoung
    2017 22ND ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2017, : 696 - 701
  • [25] A Review of Convolutional Neural Networks Hardware Accelerators for AIoT Edge Computing
    Wu, Fei
    Zhao, Neng
    Liu, Ye
    Chang, Liang
    Zhou, Liang
    Zhou, Jun
    2021 6TH INTERNATIONAL CONFERENCE ON UK-CHINA EMERGING TECHNOLOGIES (UCET 2021), 2021, : 71 - 76
  • [26] Hardware Spiking Neural Networks with Pair-Based STDP Using Stochastic Computing
    Junxiu Liu
    Yanhu Wang
    Yuling Luo
    Shunsheng Zhang
    Dong Jiang
    Yifan Hua
    Sheng Qin
    Su Yang
    Neural Processing Letters, 2023, 55 : 7155 - 7173
  • [27] Hardware Spiking Neural Networks with Pair-Based STDP Using Stochastic Computing
    Liu, Junxiu
    Wang, Yanhu
    Luo, Yuling
    Zhang, Shunsheng
    Jiang, Dong
    Hua, Yifan
    Qin, Sheng
    Yang, Su
    NEURAL PROCESSING LETTERS, 2023, 55 (06) : 7155 - 7173
  • [28] A sparsity-based stochastic pooling mechanism for deep convolutional neural networks
    Song, Zhenhua
    Liu, Yan
    Song, Rong
    Chen, Zhenguang
    Yang, Jianyong
    Zhang, Chao
    Jiang, Qing
    NEURAL NETWORKS, 2018, 105 : 340 - 345
  • [29] Optimizing nonlinear activation function for convolutional neural networks
    Munender Varshney
    Pravendra Singh
    Signal, Image and Video Processing, 2021, 15 : 1323 - 1330
  • [30] Optimizing nonlinear activation function for convolutional neural networks
    Varshney, Munender
    Singh, Pravendra
    SIGNAL IMAGE AND VIDEO PROCESSING, 2021, 15 (06) : 1323 - 1330