CyNAPSE: A Low-power Reconfigurable Neural Inference Accelerator for Spiking Neural Networks

被引:2
|
作者
Saha, Saunak [1 ]
Duwe, Henry [1 ]
Zambreno, Joseph [1 ]
机构
[1] Iowa State Univ, Dept Elect & Comp Engn, Ames, IA 50011 USA
基金
美国国家科学基金会;
关键词
Neuromorphic; Spiking neural networks; Reconfigurable; Accelerator; Memory; Caching; Leakage; Energy efficiency; PROCESSOR; MODEL; ARCHITECTURE;
D O I
10.1007/s11265-020-01546-x
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
While neural network models keep scaling in depth and computational requirements, biologically accurate models are becoming more interesting for low-cost inference. Coupled with the need to bring more computation to the edge in resource-constrained embedded and IoT devices, specialized ultra-low power accelerators for spiking neural networks are being developed. Having a large variance in the models employed in these networks, these accelerators need to be flexible, user-configurable, performant and energy efficient. In this paper, we describe CyNAPSE, a fully digital accelerator designed to emulate neural dynamics of diverse spiking networks. Since the use case of our implementation is primarily concerned with energy efficiency, we take a closer look at the factors that could improve its energy consumption. We observe that while majority of its dynamic power consumption can be credited to memory traffic, its on-chip components suffer greatly from static leakage. Given that the event-driven spike processing algorithm is naturally memory-intensive and has a large number of idle processing elements, it makes sense to tackle each of these problems towards a more efficient hardware implementation. With a diverse set of network benchmarks, we incorporate a detailed study of memory patterns that ultimately informs our choice of an application-specific network-adaptive memory management strategy to reduce dynamic power consumption of the chip. Subsequently, we also propose and evaluate a leakage mitigation strategy for runtime control of idle power. Using both the RTL implementation and a software simulation of CyNAPSE, we measure the relative benefits of these undertakings. Results show that our adaptive memory management policy results in up to 22% more reduction in dynamic power consumption compared to conventional policies. The runtime leakage mitigation techniques show that up to 99.92% and at least 14% savings in leakage energy consumption is achievable in CyNAPSE hardware modules.
引用
收藏
页码:907 / 929
页数:23
相关论文
共 50 条
  • [21] Neuron Circuits for Low-Power Spiking Neural Networks Using Time-To-First-Spike Encoding
    Oh, Seongbin
    Kwon, Dongseok
    Yeom, Gyuho
    Kang, Won-Mook
    Lee, Soochang
    Woo, Sung Yun
    Kim, Jaehyeon
    Lee, Jong-Ho
    IEEE ACCESS, 2022, 10 : 24444 - 24455
  • [22] A dynamic decoder with speculative termination for low latency inference in spiking neural networks
    Yang, Yuxuan
    Xuan, Zihao
    Chen, Song
    Kang, Yi
    NEUROCOMPUTING, 2025, 624
  • [23] Quantisation and pooling method for low-inference-latency spiking neural networks
    Lin, Zhitao
    Shen, Juncheng
    Ma, De
    Meng, Jianyi
    ELECTRONICS LETTERS, 2017, 53 (20) : 1347 - 1348
  • [24] A Low-power Low-noise Reconfigurable Bandwidth BiCMOS Neural Amplifier
    Tasneem, Nishat T.
    Mahbub, Ifana
    PROCEEDINGS OF THE 2018 IEEE 13TH DALLAS CIRCUITS AND SYSTEMS CONFERENCE (DCAS), 2018,
  • [25] LHC: A Low-power Heterogeneous Computing Method on Neural Network Accelerator
    Liu, Fangxin
    Xie, Kunpeng
    Gong, Cheng
    Liu, Shusheng
    Lu, Ye
    Li, Tao
    2019 IEEE 25TH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS (ICPADS), 2019, : 326 - 334
  • [26] Hypercolumn Sparsification for Low-Power Convolutional Neural Networks
    Pilly, Praveen K.
    Stepp, Nigel D.
    Liapis, Yannis
    Payton, David W.
    Srinivasa, Narayan
    ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2019, 15 (02)
  • [27] QUENN: QUantization Engine for low-power Neural Networks
    de Prado, Miguel
    Denna, Maurizio
    Benini, Luca
    Pazos, Nuria
    2018 ACM INTERNATIONAL CONFERENCE ON COMPUTING FRONTIERS, 2018, : 36 - 44
  • [28] A low-power analog implementation of cellular neural networks
    Anguita, M
    Pelayo, FJ
    Fernandez, FJ
    Prieto, A
    FROM NATURAL TO ARTIFICIAL NEURAL COMPUTATION, 1995, 930 : 736 - 743
  • [29] A Low-Voltage, Low-Power Reconfigurable Current-Mode Softmax Circuit for Analog Neural Networks
    Vatalaro, Massimo
    Moposita, Tatiana
    Strangio, Sebastiano
    Trojman, Lionel
    Vladimirescu, Andrei
    Lanuzza, Marco
    Crupi, Felice
    ELECTRONICS, 2021, 10 (09)
  • [30] An FPGA implementation of Bayesian inference with spiking neural networks
    Li, Haoran
    Wan, Bo
    Fang, Ying
    Li, Qifeng
    Liu, Jian K.
    An, Lingling
    FRONTIERS IN NEUROSCIENCE, 2024, 17