Hardware-Driven Nonlinear Activation for Stochastic Computing Based Deep Convolutional Neural Networks

被引:0
|
作者
Li, Ji [1 ]
Yuan, Zihao [1 ]
Li, Zhe [2 ]
Ding, Caiwen [2 ]
Ren, Ao [2 ]
Qiu, Qinru [2 ]
Draper, Jeffrey [1 ,3 ]
Wang, Yanzhi [2 ]
机构
[1] Univ Southern Calif, Dept Elect Engn, Los Angeles, CA 90007 USA
[2] Syracuse Univ, Coll Engn & Comp Sci, Syracuse, NY USA
[3] Univ Southern Calif, Informat Sci Inst, Marina Del Rey, CA 90292 USA
关键词
Deep Convolutional Neural Networks; Stochastic Computing; Deep Learning; Activation Function;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, Deep Convolutional Neural Networks (DCNNs) have made unprecedented progress, achieving the accuracy close to, or even better than human-level perception in various tasks. There is a timely need to map the latest software DCNNs to application-specific hardware, in order to achieve orders of magnitude improvement in performance, energy efficiency and compactness. Stochastic Computing (SC), as a low-cost alternative to the conventional binary computing paradigm, has the potential to enable massively parallel and highly scalable hardware implementation of DCNNs. One major challenge in SC based DCNNs is designing accurate nonlinear activation functions, which have a significant impact on the network-level accuracy but cannot be implemented accurately by existing SC computing blocks. In this paper, we design and optimize SC based neurons, and we propose highly accurate activation designs for the three most frequently used activation functions in software DCNNs, i.e, hyperbolic tangent, logistic, and rectified linear units. Experimental results on LeNet-5 using MNIST dataset demonstrate that compared with a binary ASIC hardware DCNN, the DCNN with the proposed SC neurons can achieve up to 61X, 151X, and 2X improvement in terms of area, power, and energy, respectively, at the cost of small precision degradation. In addition, the SC approach achieves up to 21X and 41X of the area, 41X and 72X of the power, and 198200X and 96443X of the energy, compared with CPU and GPU approaches, respectively, while the error is increased by less than 3.07%. ReLU activation is suggested for future SC based DCNNs considering its superior performance under a small bit stream length.
引用
收藏
页码:1230 / 1236
页数:7
相关论文
共 50 条
  • [31] DPS: Dynamic Precision Scaling for Stochastic Computing -based Deep Neural Networks
    Sim, Hyeonuk
    Kenzhegulov, Saken
    Lee, Jongeun
    2018 55TH ACM/ESDA/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2018,
  • [32] Efficient Hardware Design of Convolutional Neural Networks for Accelerated Deep Learning
    Khalil, Kasem
    Khan, Md Rahat
    Bayoumi, Magdy
    Sherif, Ahmed
    2024 IEEE 67TH INTERNATIONAL MIDWEST SYMPOSIUM ON CIRCUITS AND SYSTEMS, MWSCAS 2024, 2024, : 1075 - 1079
  • [33] Low power & mobile hardware accelerators for deep convolutional neural networks
    Scanlan, Anthony G.
    INTEGRATION-THE VLSI JOURNAL, 2019, 65 : 110 - 127
  • [34] Optimization for deep convolutional neural network of stochastic computing on MLC-PCM-based system
    Wang, Zhaoyang
    Jia, Zhiping
    Shen, Zhaoyan
    Zhao, Yijun
    Chen, Renhai
    MICROPROCESSORS AND MICROSYSTEMS, 2022, 90
  • [35] Optimization for deep convolutional neural network of stochastic computing on MLC-PCM-based system
    Wang, Zhaoyang
    Jia, Zhiping
    Shen, Zhaoyan
    Zhao, Yijun
    Chen, Renhai
    Microprocessors and Microsystems, 2022, 90
  • [36] Evaluation of the Impact of Random Computing Hardware Faults on the Performance of Convolutional Neural Networks
    Valiev, Emil
    Morozov, Andrey
    Beyer, Michael
    Yusupova, Nafisa
    Janschek, Klaus
    PROCEEDINGS OF THE 8TH SCIENTIFIC CONFERENCE ON INFORMATION TECHNOLOGIES FOR INTELLIGENT DECISION MAKING SUPPORT (ITIDS 2020), 2020, 174 : 307 - 312
  • [37] A hardware-efficient computing engine for FPGA-based deep convolutional neural network accelerator
    Li, Xueming
    Huang, Hongmin
    Chen, Taosheng
    Gao, Huaien
    Hu, Xianghong
    Xiong, Xiaoming
    MICROELECTRONICS JOURNAL, 2022, 128
  • [38] Optimizing Stochastic Computing for Low Latency Inference of Convolutional Neural Networks
    Chen, Zhiyuan
    Ma, Yufei
    Wang, Zhongfeng
    2020 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED-DESIGN (ICCAD), 2020,
  • [39] SkippyNN: An Embedded Stochastic-Computing Accelerator for Convolutional Neural Networks
    Hojabr, Reza
    Givaki, Kamyar
    Tayaranian, S. M. Reza
    Esfahanian, Parsa
    Khonsari, Ahmad
    Rahmati, Dara
    Najafi, M. Hassan
    PROCEEDINGS OF THE 2019 56TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2019,
  • [40] Hardware-aware neural architecture search for stochastic computing-based neural networks on tiny devices
    Song, Yuhong
    Sha, Edwin Hsing-Mean
    Zhuge, Qingfeng
    Xu, Rui
    Xu, Xiaowei
    Li, Bingzhe
    Yang, Lei
    JOURNAL OF SYSTEMS ARCHITECTURE, 2023, 135