Optimizing Stochastic Computing for Low Latency Inference of Convolutional Neural Networks

被引:2
|
作者
Chen, Zhiyuan [1 ]
Ma, Yufei [1 ]
Wang, Zhongfeng [1 ]
机构
[1] Nanjing Univ, Sch Elect Sci & Engn, Nanjing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1145/3400302.3415697
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The appealing property of low area, low power, flexible precision, and high bit error tolerance has made Stochastic Computing (SC) a promising alternative to conventional binary arithmetic for many computation intensive tasks, e.g., convolutional neural networks (CNNs). However, to relieve the intrinsic fluctuation noise in SC, long bit stream is normally required in SC-based CNN accelerators to achieve satisfactory accuracy, which leads to extortionate latency. Although the bit parallel structure of a SC multiplier has been proposed to reduce latency, the resulting extra overhead still considerably degrade the overall efficiency of SC. In this paper, we optimize both the micro-architecture of SC multiply-and-accumulate (MAC) unit and the overall acceleration scheme of CNN accelerator to favor SC. An optimized and scalable SC-MAC unit, which fully utilizes the property of low-discrepancy bit stream, is proposed with adjustable parameters to reduce the latency with minor area increase. For the overall accelerator, the parallel dimensions of SC-based MAC array are extended to reuse hardware resources and improve throughput, since the judiciously chosen loop unrolling strategy can better benefit SC operations. The proposed CNN accelerator with extended SC-MAC array is synthesized and demonstrated using TSMC 28nm CMOS on several representative CNNs, which gains 2x performance speedup, 2.8x energy savings and 15% area reduction compared to state-of-the-art SC based CNN accelerator.
引用
收藏
页数:7
相关论文
共 50 条
  • [1] A Low-Latency Inference of Randomly Wired Convolutional Neural Networks on an FPGA
    Kuramochi, Ryosuke
    Nakahara, Hiroki
    [J]. IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2021, E104D (12) : 2068 - 2077
  • [2] Hartley Stochastic Computing For Convolutional Neural Networks
    Mozafari, S. H.
    Clark, J. J.
    Gross, W. J.
    Meyer, B. H.
    [J]. 2021 IEEE WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS 2021), 2021, : 235 - 240
  • [3] Stochastic Dividers for Low Latency Neural Networks
    Liu, Shanshan
    Tang, Xiaochen
    Niknia, Farzad
    Reviriego, Pedro
    Liu, Weiqiang
    Louri, Ahmed
    Lombardi, Fabrizio
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2021, 68 (10) : 4102 - 4115
  • [4] Accurate and Efficient Stochastic Computing Hardware for Convolutional Neural Networks
    Yu, Joonsang
    Kim, Kyounghoon
    Lee, Jongeun
    Choi, Kiyoung
    [J]. 2017 IEEE 35TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD), 2017, : 105 - 112
  • [5] Scalable Stochastic-Computing Accelerator for Convolutional Neural Networks
    Sim, Hyeonuk
    Dong Nguyen
    Lee, Jongeun
    Choi, Kiyoung
    [J]. 2017 22ND ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2017, : 696 - 701
  • [6] Embedded GPU Cluster Computing Framework for Inference of Convolutional Neural Networks
    Kain, Evan
    Wildenstein, Diego
    Pineda, Andrew C.
    [J]. 2019 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE (HPEC), 2019,
  • [7] Accelerating Inference of Convolutional Neural Networks Using In-memory Computing
    Dazzi, Martino
    Sebastian, Abu
    Benini, Luca
    Eleftheriou, Evangelos
    [J]. FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2021, 15
  • [8] Optimizing Convolutional Neural Networks for low-resource devices
    Rusu, Cosmin-Ionut
    Czibula, Gabriela
    [J]. 2018 IEEE 14TH INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTER COMMUNICATION AND PROCESSING (ICCP), 2018, : 245 - 252
  • [9] Towards Acceleration of Deep Convolutional Neural Networks using Stochastic Computing
    Li, Ji
    Ren, Ao
    Li, Zhe
    Ding, Caiwen
    Yuan, Bo
    Qiu, Qinru
    Wang, Yanzhi
    [J]. 2017 22ND ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2017, : 115 - 120
  • [10] SkippyNN: An Embedded Stochastic-Computing Accelerator for Convolutional Neural Networks
    Hojabr, Reza
    Givaki, Kamyar
    Tayaranian, S. M. Reza
    Esfahanian, Parsa
    Khonsari, Ahmad
    Rahmati, Dara
    Najafi, M. Hassan
    [J]. PROCEEDINGS OF THE 2019 56TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2019,