Can Deep Neural Networks be Converted to Ultra Low-Latency Spiking Neural Networks?

被引:0
|
作者
Datta, Gourav [1 ]
Beerel, Peter A. [1 ]
机构
[1] Univ Southern Calif, Ming Hsieh Dept Elect & Comp Engn, Los Angeles, CA 90089 USA
关键词
SNN; DNN; neuromorphic; FLOPs; surrogate gradient learning;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Spiking neural networks (SNNs), that operate via binary spikes distributed over time, have emerged as a promising energy efficient ML paradigm for resource-constrained devices. However, the current state-of-the-art (SOTA) SNNs require multiple time steps for acceptable inference accuracy, increasing spiking activity and, consequently, energy consumption. SOTA training strategies for SNNs involve conversion from a non-spiking deep neural network (DNN). In this paper, we determine that SOTA conversion strategies cannot yield ultra low latency because they incorrectly assume that the DNN and SNN pre-activation values are uniformly distributed. We propose a new training algorithm that accurately captures these distributions, minimizing the error between the DNN and converted SNN. The resulting SNNs have ultra low latency and high activation sparsity, yielding significant improvements in compute efficiency. In particular, we evaluate our framework on image recognition tasks from CIFAR-10 and CIFAR-100 datasets on several VGG and ResNet architectures. We obtain top-1 accuracy of 64.19% with only 2 time steps on the CIFAR-100 dataset with similar to 159.2x lower compute energy compared to an iso-architecture standard DNN. Compared to other SOTA SNN models, our models perform inference 2.5-8x faster (i.e., with fewer time steps).
引用
收藏
页码:718 / 723
页数:6
相关论文
共 50 条
  • [1] Optimized Potential Initialization for Low-Latency Spiking Neural Networks
    Bu, Tong
    Ding, Jianhao
    Yu, Zhaofei
    Huang, Tiejun
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 11 - 20
  • [2] Constrain Bias Addition to Train Low-Latency Spiking Neural Networks
    Lin, Ranxi
    Dai, Benzhe
    Zhao, Yingkai
    Chen, Gang
    Lu, Huaxiang
    BRAIN SCIENCES, 2023, 13 (02)
  • [3] Revisiting Batch Normalization for Training Low-Latency Deep Spiking Neural Networks From Scratch
    Kim, Youngeun
    Panda, Priyadarshini
    FRONTIERS IN NEUROSCIENCE, 2021, 15
  • [4] Highway Connection for Low-Latency and High-Accuracy Spiking Neural Networks
    Zhang, Anguo
    Wu, Junyi
    Li, Xiumin
    Li, Hung Chun
    Gao, Yueming
    Pun, Sio Hang
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2023, 70 (12) : 4579 - 4583
  • [5] Spatio-Temporal Pruning and Quantization for Low-latency Spiking Neural Networks
    Chowdhury, Sayeed Shafayet
    Garg, Isha
    Roy, Kaushik
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [6] LOW-LATENCY SOUND SOURCE SEPARATION USING DEEP NEURAL NETWORKS
    Naithani, Gaurav
    Parascandolo, Giambattista
    Barker, Tom
    Pontoppidan, Niels Henrik
    Virtanen, Tuomas
    2016 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP), 2016, : 272 - 276
  • [7] Energy efficient and low-latency spiking neural networks on embedded microcontrollers through spiking activity tuning
    Francesco Barchi
    Emanuele Parisi
    Luca Zanatta
    Andrea Bartolini
    Andrea Acquaviva
    Neural Computing and Applications, 2024, 36 (30) : 18897 - 18917
  • [8] Amortized Neural Networks for Low-Latency Speech Recognition
    Macoskey, Jonathan
    Strimel, Grant P.
    Su, Jinru
    Rastrow, Ariya
    INTERSPEECH 2021, 2021, : 4558 - 4562
  • [9] Training High-Performance Low-Latency Spiking Neural Networks by Differentiation on Spike Representation
    Meng, Qingyan
    Xiao, Mingqing
    Yan, Shen
    Wang, Yisen
    Lin, Zhouchen
    Luo, Zhi-Quan
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 12434 - 12443
  • [10] Shrinking Your TimeStep: Towards Low-Latency Neuromorphic Object Recognition with Spiking Neural Networks
    Ding, Yongqi
    Zuo, Lin
    Jing, Mengmeng
    He, Pei
    Xiao, Yongjun
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 10, 2024, : 11811 - 11819