Quantisation and pooling method for low-inference-latency spiking neural networks

被引:10
|
作者
Lin, Zhitao [1 ]
Shen, Juncheng [1 ]
Ma, De [2 ]
Meng, Jianyi [3 ]
机构
[1] Zhejiang Univ, Inst VLSI Design, Hangzhou, Zhejiang, Peoples R China
[2] Zhejiang Univ, Coll Comp Sci, Hangzhou, Zhejiang, Peoples R China
[3] Fudan Univ, State Key Lab ASIC & Syst, Shanghai, Peoples R China
关键词
neural nets; object recognition; real-time recognition tasks; CIFAR10; MNIST; spiking neurons; convolutional layers; pooling function; retraining; layer-wise quantisation method; DNN; deep neural network; SNN; low-inference-latency spiking neural networks; pooling method;
D O I
10.1049/el.2017.2219
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Spiking neural network (SNN) that converted from conventional deep neural network (DNN) has shown great potential as a solution for fast and efficient recognition. A layer-wise quantisation method based on retraining is proposed to quantise the activation of DNN, which reduces the number of time steps required by converted SNN to achieve minimal accuracy loss. Pooling function is incorporated into convolutional layers to reduce at most 20% of spiking neurons. The converted SNNs achieved 99.15% accuracy on MNIST and 82.9% on CIFAR10 by only seven time steps, and only 10-40% of spikes need to be processed compared with networks using traditional algorithms. The experimental results show that the proposed methods are able to build hardware-friendly SNNs with ultra-low-inference latency.
引用
收藏
页码:1347 / 1348
页数:2
相关论文
共 50 条
  • [1] Can Deep Neural Networks be Converted to Ultra Low-Latency Spiking Neural Networks?
    Datta, Gourav
    Beerel, Peter A.
    PROCEEDINGS OF THE 2022 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2022), 2022, : 718 - 723
  • [2] Optimized Potential Initialization for Low-Latency Spiking Neural Networks
    Bu, Tong
    Ding, Jianhao
    Yu, Zhaofei
    Huang, Tiejun
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 11 - 20
  • [3] CyNAPSE: A Low-power Reconfigurable Neural Inference Accelerator for Spiking Neural Networks
    Saunak Saha
    Henry Duwe
    Joseph Zambreno
    Journal of Signal Processing Systems, 2020, 92 : 907 - 929
  • [4] Constrain Bias Addition to Train Low-Latency Spiking Neural Networks
    Lin, Ranxi
    Dai, Benzhe
    Zhao, Yingkai
    Chen, Gang
    Lu, Huaxiang
    BRAIN SCIENCES, 2023, 13 (02)
  • [5] CyNAPSE: A Low-power Reconfigurable Neural Inference Accelerator for Spiking Neural Networks
    Saha, Saunak
    Duwe, Henry
    Zambreno, Joseph
    JOURNAL OF SIGNAL PROCESSING SYSTEMS FOR SIGNAL IMAGE AND VIDEO TECHNOLOGY, 2020, 92 (09): : 907 - 929
  • [6] A lightweight Max-Pooling method and architecture for Deep Spiking Convolutional Neural Networks
    Duy-Anh Nguyen
    Xuan-Tu Tran
    Dang, Khanh N.
    Iacopi, Francesca
    APCCAS 2020: PROCEEDINGS OF THE 2020 IEEE ASIA PACIFIC CONFERENCE ON CIRCUITS AND SYSTEMS (APCCAS 2020), 2020, : 209 - 212
  • [7] Effective Plug-Ins for Reducing Inference-Latency of Spiking Convolutional Neural Networks During Inference Phase
    Chen, Xuan
    Yuan, Xiaopeng
    Fu, Gaoming
    Luo, Yuanyong
    Yue, Tao
    Yan, Feng
    Wang, Yuxuan
    Pan, Hongbing
    FRONTIERS IN COMPUTATIONAL NEUROSCIENCE, 2021, 15
  • [8] Optimizing Stochastic Computing for Low Latency Inference of Convolutional Neural Networks
    Chen, Zhiyuan
    Ma, Yufei
    Wang, Zhongfeng
    2020 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED-DESIGN (ICCAD), 2020,
  • [9] Bayesian Inference Accelerator for Spiking Neural Networks
    Katti, Prabodh
    Nimbekar, Anagha
    Li, Chen
    Acharyya, Amit
    Al-Hashimi, Bashir M.
    Rajendran, Bipin
    2024 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, ISCAS 2024, 2024,
  • [10] Energy efficient and low-latency spiking neural networks on embedded microcontrollers through spiking activity tuning
    Francesco Barchi
    Emanuele Parisi
    Luca Zanatta
    Andrea Bartolini
    Andrea Acquaviva
    Neural Computing and Applications, 2024, 36 (30) : 18897 - 18917