Can Deep Neural Networks be Converted to Ultra Low-Latency Spiking Neural Networks?

被引:0
|
作者
Datta, Gourav [1 ]
Beerel, Peter A. [1 ]
机构
[1] Univ Southern Calif, Ming Hsieh Dept Elect & Comp Engn, Los Angeles, CA 90089 USA
关键词
SNN; DNN; neuromorphic; FLOPs; surrogate gradient learning;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Spiking neural networks (SNNs), that operate via binary spikes distributed over time, have emerged as a promising energy efficient ML paradigm for resource-constrained devices. However, the current state-of-the-art (SOTA) SNNs require multiple time steps for acceptable inference accuracy, increasing spiking activity and, consequently, energy consumption. SOTA training strategies for SNNs involve conversion from a non-spiking deep neural network (DNN). In this paper, we determine that SOTA conversion strategies cannot yield ultra low latency because they incorrectly assume that the DNN and SNN pre-activation values are uniformly distributed. We propose a new training algorithm that accurately captures these distributions, minimizing the error between the DNN and converted SNN. The resulting SNNs have ultra low latency and high activation sparsity, yielding significant improvements in compute efficiency. In particular, we evaluate our framework on image recognition tasks from CIFAR-10 and CIFAR-100 datasets on several VGG and ResNet architectures. We obtain top-1 accuracy of 64.19% with only 2 time steps on the CIFAR-100 dataset with similar to 159.2x lower compute energy compared to an iso-architecture standard DNN. Compared to other SOTA SNN models, our models perform inference 2.5-8x faster (i.e., with fewer time steps).
引用
收藏
页码:718 / 723
页数:6
相关论文
共 50 条
  • [21] A LOW-LATENCY SPARSE-WINOGRAD ACCELERATOR FOR CONVOLUTIONAL NEURAL NETWORKS
    Wang, Haonan
    Liu, Wenjian
    Xu, Tianyi
    Lin, Jun
    Wang, Zhongfeng
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 1448 - 1452
  • [22] A Low-Latency Inference of Randomly Wired Convolutional Neural Networks on an FPGA
    Kuramochi, Ryosuke
    Nakahara, Hiroki
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2021, E104D (12) : 2068 - 2077
  • [23] DSNNs: learning transfer from deep neural networks to spiking neural networks
    Zhang L.
    Du Z.
    Li L.
    Chen Y.
    High Technology Letters, 2020, 26 (02): : 136 - 144
  • [24] DSNNs:learning transfer from deep neural networks to spiking neural networks
    张磊
    Du Zidong
    Li Ling
    Chen Yunji
    High Technology Letters, 2020, 26 (02) : 136 - 144
  • [25] Converting Artificial Neural Networks to Ultralow-Latency Spiking Neural Networks for Action Recognition
    You, Hong
    Zhong, Xian
    Liu, Wenxuan
    Wei, Qi
    Huang, Wenxin
    Yu, Zhaofei
    Huang, Tiejun
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2024, 16 (04) : 1533 - 1545
  • [26] DCT-SNN: Using DCT to Distribute Spatial Information over Time for Low-Latency Spiking Neural Networks
    Garg, Isha
    Chowdhury, Sayeed Shafayet
    Roy, Kaushik
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 4651 - 4660
  • [27] Quantisation and pooling method for low-inference-latency spiking neural networks
    Lin, Zhitao
    Shen, Juncheng
    Ma, De
    Meng, Jianyi
    ELECTRONICS LETTERS, 2017, 53 (20) : 1347 - 1348
  • [28] Training Low-Latency Spiking Neural Network through Knowledge Distillation
    Takuya, Sugahara
    Zhang, Renyuan
    Nakashima, Yasuhiko
    2021 IEEE COOL CHIPS 24: IEEE SYMPOSIUM IN LOW-POWER AND HIGH-SPEED CHIPS, 2021,
  • [29] Ultra-low latency spiking neural networks with spatio-temporal compression and synaptic convolutional block
    Xu, Changqing
    Liu, Yi
    Yang, Yintang
    NEUROCOMPUTING, 2023, 550
  • [30] Direct Training via Backpropagation for Ultra-Low-Latency Spiking Neural Networks with Multi-Threshold
    Xu, Changqing
    Liu, Yi
    Chen, Dongdong
    Yang, Yintang
    SYMMETRY-BASEL, 2022, 14 (09):