SC-DCNN: Highly-Scalable Deep Convolutional Neural Network using Stochastic Computing

被引:41
|
作者
Ren, Ao [1 ]
Li, Zhe [1 ]
Ding, Caiwen [1 ]
Qiu, Qinru [1 ]
Wang, Yanzhi [1 ]
Li, Ji [2 ]
Qian, Xuehai [2 ]
Yuan, Bo [3 ]
机构
[1] Syracuse Univ, Dept Elect Engn & Comp Sci, Syracuse, NY 13244 USA
[2] Univ Southern Calif, Dept Elect Engn, Los Angeles, CA 90089 USA
[3] CUNY City Coll, Dept Elect Engn, New York, NY 10031 USA
关键词
IMPLEMENTATION;
D O I
10.1145/3093336.3037746
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
With the recent advance of wearable devices and Internet of Things (IoTs), it becomes attractive to implement the Deep Convolutional Neural Networks (DCNNs) in embedded and portable systems. Currently, executing the software-based DCNNs requires high-performance servers, restricting the widespread deployment on embedded and mobile IoT devices. To overcome this obstacle, considerable research efforts have been made to develop highly-parallel and specialized DCNN accelerators using GPGPUs, FPGAs or ASICs. Stochastic Computing (SC), which uses a bit-stream to represent a number within [-1, 1] by counting the number of ones in the bit-stream, has high potential for implementing DCNNs with high scalability and ultra-low hardware footprint. Since multiplications and additions can be calculated using AND gates and multiplexers in SC, significant reductions in power (energy) and hardware footprint can be achieved compared to the conventional binary arithmetic implementations. The tremendous savings in power (energy) and hardware resources allow immense design space for enhancing scalability and robustness for hardware DCNNs. This paper presents SC-DCNN, the first comprehensive design and optimization framework of SC-based DCNNs, using a bottom-up approach. We first present the designs of function blocks that perform the basic operations in DCNN, including inner product, pooling, and activation function. Then we propose four designs of feature extraction blocks, which are in charge of extracting features from input feature maps, by connecting different basic function blocks with joint optimization. Moreover, the efficient weight storage methods are proposed to reduce the area and power (energy) consumption. Putting all together, with feature extraction blocks carefully selected, SC-DCNN is holistically optimized to minimize area and power (energy) consumption while maintaining high network accuracy. Experimental results demonstrate that the LeNet5 implemented in SCDCNN consumes only 17 mm(2) area and 1.53 W power, achieves throughput of 781250 images/s, area efficiency of 45946 images/s/mm(2), and energy efficiency of 510734 images/J.
引用
收藏
页码:405 / 418
页数:14
相关论文
共 50 条
  • [31] Hyperspectral image classification using deep convolutional neural network and stochastic relaxation labeling
    Singh, Manoj K.
    Mohan, Shashank
    Kumar, Brajesh
    JOURNAL OF APPLIED REMOTE SENSING, 2021, 15 (04)
  • [32] Applying Deep Convolutional Neural Network (DCNN) Algorithm in the Cloud Autonomous Vehicles Traffic Model
    Ramakrishnan, Dhaya
    Radhakrishnan, Kanthavel
    INTERNATIONAL ARAB JOURNAL OF INFORMATION TECHNOLOGY, 2022, 19 (02) : 186 - 194
  • [33] A Deep Convolutional Neural Network (DCNN) with Fine Tuned Hyper Parameters for Skin Cancer Classification
    Shaik, Abdul Rahaman
    Pullagura, Rajesh Kumar
    Traitement du Signal, 2024, 41 (05) : 2623 - 2633
  • [34] DCNN: Deep Convolutional Neural Network With XAI for Efficient Detection of Specific Language Impairment in Children
    Md Hasib, Khan
    Mridha, M. F.
    Mehedi, Md Humaion Kabir
    Faruk, Kazi Omar
    Muna, Rabeya Khatun
    Iqbal, Shahriar
    Islam, Md Rashedul
    Watanobe, Yutaka
    IEEE ACCESS, 2024, 12 : 101660 - 101678
  • [35] Multi-class Classification of Impulse and Non-impulse Sounds Using Deep Convolutional Neural Network (DCNN)
    Abayomi-Alli, Adebayo
    Abayomi-Alli, Olusola
    Vipperman, Jeffrey
    Odusami, Modupe
    Misra, Sanjay
    COMPUTATIONAL SCIENCE AND ITS APPLICATIONS - ICCSA 2019, PT V: 19TH INTERNATIONAL CONFERENCE, SAINT PETERSBURG, RUSSIA, JULY 14, 2019, PROCEEDINGS, PART V, 2019, 11623 : 359 - 371
  • [36] EPM-DCNN: Earthquake Prediction Models Using Deep Convolutional Neural Networks
    Shan, Weifeng
    Zhang, Mingjie
    Wang, Maofa
    Chen, Huiling
    Zhang, Ruilei
    Yang, Guangze
    Tang, Yixiang
    Teng, Yuntian
    Chen, Jun
    BULLETIN OF THE SEISMOLOGICAL SOCIETY OF AMERICA, 2022, 112 (06) : 2933 - 2945
  • [37] Ascend: a Scalable and Unified Architecture for Ubiquitous Deep Neural Network Computing
    Liao, Heng
    Tu, Jiajin
    Xia, Jing
    Liu, Hu
    Zhou, Xiping
    Yuan, Honghui
    Hu, Yuxing
    2021 27TH IEEE INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE COMPUTER ARCHITECTURE (HPCA 2021), 2021, : 789 - 801
  • [38] A Low-cost Implementation Method on Deep Neural Network Using Stochastic Computing
    Dong, Ya
    Xiong, Xingzhong
    Li, Tianyu
    Zhang, Lin
    Chen, Jienan
    SECOND IYSF ACADEMIC SYMPOSIUM ON ARTIFICIAL INTELLIGENCE AND COMPUTER ENGINEERING, 2021, 12079
  • [39] Normalization and dropout for stochastic computing-based deep convolutional neural networks
    Li, Ji
    Yuan, Zihao
    Li, Zhe
    Ren, Ao
    Ding, Caiwen
    Draper, Jeffrey
    Nazarian, Shahin
    Qiu, Qinru
    Yuan, Bo
    Wang, Yanzhi
    INTEGRATION-THE VLSI JOURNAL, 2019, 65 : 395 - 403
  • [40] Softmax Regression Design for Stochastic Computing Based Deep Convolutional Neural Networks
    Yuan, Zihao
    Li, Ji
    Li, Zhe
    Ding, Caiwen
    Ren, Ao
    Yuan, Bo
    Qiu, Qinru
    Draper, Jeffrey
    Wang, Yanzhi
    PROCEEDINGS OF THE GREAT LAKES SYMPOSIUM ON VLSI 2017 (GLSVLSI' 17), 2017, : 467 - 470