A low-power, high-accuracy with fully on-chip ternary weight hardware architecture for Deep Spiking Neural Networks

被引:5
|
作者
Duy-Anh Nguyen [1 ,2 ]
Xuan-Tu Tran [1 ]
Dang, Khanh N. [3 ]
Iacopi, Francesca [4 ]
机构
[1] Vietnam Natl Univ Hanoi VNU, VNU Informat Technol Inst, Hanoi 123106, Vietnam
[2] VNU UET, JTIRC, Hanoi, Vietnam
[3] Vietnam Natl Univ Hanoi VNU, VNU Key Lab Smart Integrated Syst SISLAB, VNU UET, Hanoi 123106, Vietnam
[4] Univ Technol Sydney, 15 Broadway, Ultimo, NSW 2007, Australia
关键词
Deep Spiking Neural Network; Neuromorphic; Ternary-weight quantization; Hardware implementation; EFFICIENT;
D O I
10.1016/j.micpro.2022.104458
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Recently, Deep Spiking Neural Network (DSNN) has emerged as a promising neuromorphic approach for various AI-based applications, such as image classification, speech recognition, robotic control etc. on edge computing platforms. However, the state-of-the-art offline training algorithms for DSNNs are facing two major challenges. Firstly, many timesteps are required to reach comparable accuracy with traditional frame-based DNNs algorithms. Secondly, extensive memory requirements for weight storage make it impossible to store all the weights on-chip for DSNNs with many layers. Thus the inference process requires continue access to expensive off-chip memory, ultimately leading to performance degradation in terms of throughput and power consumption. In this work, we propose a hardware-friendly training approach for DSNN that allows the weights to be constrained to ternary format, hence reducing the memory footprints and the energy consumption. Software simulations on MNIST and CIFAR10 datasets have shown that our training approach could reach an accuracy of 97% for MNIST (3-layer fully connected networks) and 89.71% for CIFAR10 (VGG16). To demonstrate the energy efficiency of our approach, we have proposed a neural processing module to implement our trained DSNN. When implemented as a fixed, 3-layers fully-connected system, the system has reached at energy efficiency of 74nJ/image with a classification accuracy of 97% for MNIST dataset. We have also considered a scalable design to support more complex network topologies when we integrate the neural processing module with a 3D Network-on-Chip.
引用
收藏
页数:15
相关论文
共 50 条
  • [21] Adaptive Data Compression for High-Performance Low-Power On-Chip Networks
    Jin, Yuho
    Yum, Ki Hwan
    Kim, Eun Jung
    2008 PROCEEDINGS OF THE 41ST ANNUAL IEEE/ACM INTERNATIONAL SYMPOSIUM ON MICROARCHITECTURE: MICRO-41, 2008, : 354 - +
  • [22] A Low-Latency and Low-Power Hybrid Scheme for On-Chip Networks
    Jiang, Guoyue
    Li, Zhaolin
    Wang, Fang
    Wei, Shaojun
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2015, 23 (04) : 664 - 677
  • [23] Toward High-Accuracy and Low-Latency Spiking Neural Networks With Two-Stage Optimization
    Wang, Ziming
    Zhang, Yuhao
    Lian, Shuang
    Cui, Xiaoxin
    Yan, Rui
    Tang, Huajin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, : 1 - 15
  • [24] Serial-link bus: A low-power on-chip bus architecture
    Ghoneima, M
    Ismail, Y
    Khellah, M
    Tschanz, J
    De, VV
    ICCAD-2005: INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN, DIGEST OF TECHNICAL PAPERS, 2005, : 541 - 546
  • [25] Serial-Link Bus: A Low-Power On-Chip Bus Architecture
    Ghoneima, Maged
    Ismail, Yehea
    Khellah, Muhammad M.
    Tschanz, James
    De, Vivek
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2009, 56 (09) : 2020 - 2032
  • [26] Spiking neural networks with consistent mapping relations allow high-accuracy inference
    Li, Yang
    He, Xiang
    Kong, Qingqun
    Zeng, Yi
    INFORMATION SCIENCES, 2024, 677
  • [27] Low-power on-chip bus architecture using dynamic relative delays
    Ghoneima, M
    Ismail, Y
    IEEE INTERNATIONAL SOC CONFERENCE, PROCEEDINGS, 2004, : 233 - 236
  • [28] CyNAPSE: A Low-power Reconfigurable Neural Inference Accelerator for Spiking Neural Networks
    Saunak Saha
    Henry Duwe
    Joseph Zambreno
    Journal of Signal Processing Systems, 2020, 92 : 907 - 929
  • [29] CyNAPSE: A Low-power Reconfigurable Neural Inference Accelerator for Spiking Neural Networks
    Saha, Saunak
    Duwe, Henry
    Zambreno, Joseph
    JOURNAL OF SIGNAL PROCESSING SYSTEMS FOR SIGNAL IMAGE AND VIDEO TECHNOLOGY, 2020, 92 (09): : 907 - 929
  • [30] CESP: A Low-Power High-Accuracy Time Synchronization Protocol
    Gong, Fengyuan
    Sichitiu, Mihail L.
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2016, 65 (04) : 2387 - 2396