Hardware-Driven Nonlinear Activation for Stochastic Computing Based Deep Convolutional Neural Networks

被引:0
|
作者
Li, Ji [1 ]
Yuan, Zihao [1 ]
Li, Zhe [2 ]
Ding, Caiwen [2 ]
Ren, Ao [2 ]
Qiu, Qinru [2 ]
Draper, Jeffrey [1 ,3 ]
Wang, Yanzhi [2 ]
机构
[1] Univ Southern Calif, Dept Elect Engn, Los Angeles, CA 90007 USA
[2] Syracuse Univ, Coll Engn & Comp Sci, Syracuse, NY USA
[3] Univ Southern Calif, Informat Sci Inst, Marina Del Rey, CA 90292 USA
关键词
Deep Convolutional Neural Networks; Stochastic Computing; Deep Learning; Activation Function;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, Deep Convolutional Neural Networks (DCNNs) have made unprecedented progress, achieving the accuracy close to, or even better than human-level perception in various tasks. There is a timely need to map the latest software DCNNs to application-specific hardware, in order to achieve orders of magnitude improvement in performance, energy efficiency and compactness. Stochastic Computing (SC), as a low-cost alternative to the conventional binary computing paradigm, has the potential to enable massively parallel and highly scalable hardware implementation of DCNNs. One major challenge in SC based DCNNs is designing accurate nonlinear activation functions, which have a significant impact on the network-level accuracy but cannot be implemented accurately by existing SC computing blocks. In this paper, we design and optimize SC based neurons, and we propose highly accurate activation designs for the three most frequently used activation functions in software DCNNs, i.e, hyperbolic tangent, logistic, and rectified linear units. Experimental results on LeNet-5 using MNIST dataset demonstrate that compared with a binary ASIC hardware DCNN, the DCNN with the proposed SC neurons can achieve up to 61X, 151X, and 2X improvement in terms of area, power, and energy, respectively, at the cost of small precision degradation. In addition, the SC approach achieves up to 21X and 41X of the area, 41X and 72X of the power, and 198200X and 96443X of the energy, compared with CPU and GPU approaches, respectively, while the error is increased by less than 3.07%. ReLU activation is suggested for future SC based DCNNs considering its superior performance under a small bit stream length.
引用
收藏
页码:1230 / 1236
页数:7
相关论文
共 50 条
  • [1] Towards Budget-Driven Hardware Optimization for Deep Convolutional Neural Networks using Stochastic Computing
    Li, Zhe
    Li, Ji
    Ren, Ao
    Ding, Caiwen
    Draper, Jeffrey
    Qiu, Qinru
    Yuan, Bo
    Wang, Yanzhi
    2018 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI), 2018, : 28 - 33
  • [2] DSCNN: Hardware-Oriented Optimization for Stochastic Computing Based Deep Convolutional Neural Networks
    Li, Zhe
    Ren, Ao
    Li, Ji
    Qiu, Qinru
    Wang, Yanzhi
    Yuan, Bo
    PROCEEDINGS OF THE 34TH IEEE INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD), 2016, : 678 - 681
  • [3] Accurate and Efficient Stochastic Computing Hardware for Convolutional Neural Networks
    Yu, Joonsang
    Kim, Kyounghoon
    Lee, Jongeun
    Choi, Kiyoung
    2017 IEEE 35TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD), 2017, : 105 - 112
  • [4] An Efficient Hardware Implementation of Activation Functions Using Stochastic Computing for Deep Neural Networks
    Van-Tinh Nguyen
    Tieu-Khanh Luong
    Han Le Duc
    Van-Phuc Hoang
    2018 IEEE 12TH INTERNATIONAL SYMPOSIUM ON EMBEDDED MULTICORE/MANY-CORE SYSTEMS-ON-CHIP (MCSOC 2018), 2018, : 233 - 236
  • [5] Normalization and dropout for stochastic computing-based deep convolutional neural networks
    Li, Ji
    Yuan, Zihao
    Li, Zhe
    Ren, Ao
    Ding, Caiwen
    Draper, Jeffrey
    Nazarian, Shahin
    Qiu, Qinru
    Yuan, Bo
    Wang, Yanzhi
    INTEGRATION-THE VLSI JOURNAL, 2019, 65 : 395 - 403
  • [6] Softmax Regression Design for Stochastic Computing Based Deep Convolutional Neural Networks
    Yuan, Zihao
    Li, Ji
    Li, Zhe
    Ding, Caiwen
    Ren, Ao
    Yuan, Bo
    Qiu, Qinru
    Draper, Jeffrey
    Wang, Yanzhi
    PROCEEDINGS OF THE GREAT LAKES SYMPOSIUM ON VLSI 2017 (GLSVLSI' 17), 2017, : 467 - 470
  • [7] Fully Parallel Stochastic Computing Hardware Implementation of Convolutional Neural Networks for Edge Computing Applications
    Frasser, Christiam F.
    Linares-Serrano, Pablo
    de los Rios, Ivan Diez
    Moran, Alejandro
    Skibinsky-Gitlin, Erik S.
    Font-Rossello, Joan
    Canals, Vincent
    Roca, Miquel
    Serrano-Gotarredona, Teresa
    Rossello, Josep L.
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (12) : 10408 - 10418
  • [8] Towards Acceleration of Deep Convolutional Neural Networks using Stochastic Computing
    Li, Ji
    Ren, Ao
    Li, Zhe
    Ding, Caiwen
    Yuan, Bo
    Qiu, Qinru
    Wang, Yanzhi
    2017 22ND ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2017, : 115 - 120
  • [9] A New Stochastic Computing Multiplier with Application to Deep Convolutional Neural Networks
    Sim, Hyeonuk
    Lee, Jongeun
    PROCEEDINGS OF THE 2017 54TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2017,
  • [10] Hartley Stochastic Computing For Convolutional Neural Networks
    Mozafari, S. H.
    Clark, J. J.
    Gross, W. J.
    Meyer, B. H.
    2021 IEEE WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS 2021), 2021, : 235 - 240