A Convolutional Spiking Neural Network Accelerator with the Sparsity-aware Memory and Compressed Weights

被引:1
|
作者
Liu, Hanqing [1 ]
Cui, Xiaole [1 ]
Zhang, Sunrui [1 ]
Yin, Mingqi [1 ]
Jiang, Yuanyuan [2 ]
Cui, Xiaoxin [2 ]
机构
[1] Peking Univ, Shenzhen Grad Sch, Shenzhen, Peoples R China
[2] Peking Univ, Beijing, Peoples R China
关键词
Neuralmorphic computing; Spiking neural network accelerator; Sparse spikes; Sparse matrix compression; Field-programmable gate array; PROCESSOR;
D O I
10.1109/ASAP61560.2024.00041
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
The spiking neural network (SNN) has advantage in the edge AI applications for its spatiotemporal sparsity. The high energy efficiency is an important concern in the study of SNN accelerator designs. In this paper, a lightweight event-driven convolutional SNN accelerator that utilizes the sparsity of both the spike events and the network weights is proposed. In the event-driven mode, the proposed accelerator uses the compressed input spikes and a spike-oriented convolution data flow. An output spike compressor is also designed. To balance the computation performance and the memory space occupancy, a spike sparsity-aware memory scheme that automatically switches the spike format by a real-time monitoring strategy is designed. The compression memories and a buffer for network weights are designed to save the on-chip memory space. The accelerator prototype is verified on the Xilinx Virtex XCVU9P FPGA platform. It achieves an equivalent performance of 139.5GFLOPS on the N-MNIST dataset. Compared to the baseline using the same computational resources, the proposed accelerator can improve the inference performance, the inference energy efficiency and the memory space by 4.6, 3.6 and 1.6 times, respectively. The proposed accelerator has advantages in energy efficiency and hardware overhead compared to the previous works on the same hardware platform.
引用
收藏
页码:163 / 171
页数:9
相关论文
共 50 条
  • [21] An Energy-Efficient Accelerator with Relative-Indexing Memory for Sparse Compressed Convolutional Neural Network
    Wu, I-Chen
    Huang, Po-Tsang
    Lo, Chin-Yang
    Hwang, Wei
    2019 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2019), 2019, : 42 - 45
  • [22] Sparsity-Aware Deep Learning Accelerator Design Supporting CNN and LSTM Operations
    Hsiao, Shen-Fu
    Chang, Hsuan-Jui
    2020 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2020,
  • [23] NEM-GNN: DAC/ADC-less, Scalable, Reconfigurable, Graph and Sparsity-Aware Near-Memory Accelerator for Graph Neural Networks
    Raman, Siddhartha Raman Sundara
    John, Lizy
    Kulkarni, Jaydeep P.
    ACM TRANSACTIONS ON ARCHITECTURE AND CODE OPTIMIZATION, 2024, 21 (02)
  • [24] A Convolutional Accelerator for Neural Networks With Binary Weights
    Ardakani, Arash
    Condo, Carlo
    Gross, Warren J.
    2018 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2018,
  • [25] A Scratchpad Spiking Neural Network Accelerator
    Karakchi, Rasha
    2024 IEEE 3RD INTERNATIONAL CONFERENCE ON COMPUTING AND MACHINE INTELLIGENCE, ICMI 2024, 2024,
  • [26] Design of a Sparsity-Aware Reconfigurable Deep Learning Accelerator Supporting Various Types of Operations
    Hsiao, Shen-Fu
    Chen, Kun-Chih
    Lin, Chih-Chien
    Chang, Hsuan-Jui
    Tsai, Bo-Ching
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2020, 10 (03) : 376 - 387
  • [27] Parallax: Sparsity-aware Data Parallel Training of Deep Neural Networks
    Kim, Soojeong
    Yu, Gyeong-In
    Park, Hojin
    Cho, Sungwoo
    Jeong, Eunji
    Ha, Hyeonmin
    Lee, Sanha
    Jeong, Joo Seong
    Chun, Byung-Gon
    PROCEEDINGS OF THE FOURTEENTH EUROSYS CONFERENCE 2019 (EUROSYS '19), 2019,
  • [28] MULTIFUNCTIONAL RRAM CHIP WITH CONFIGURABILITY FOR SPARSITY-AWARE IN-MEMORY ISNG MACHINE
    Yue, Wenshuo
    Jing, Zhaokun
    Yan, Bonan
    Tao, Yaoyu
    Zhang, Teng
    Huang, Ru
    Yang, Yuchao
    CONFERENCE OF SCIENCE & TECHNOLOGY FOR INTEGRATED CIRCUITS, 2024 CSTIC, 2024,
  • [29] A Compressed Spiking Neural Network Onto a Memcapacitive In-Memory Computing Array
    Oshio, Reon
    Sugahara, Takuya
    Sawada, Atsushi
    Kimura, Mutsumi
    Zhang, Renyuan
    Nakashima, Yasuhiko
    IEEE MICRO, 2024, 44 (01) : 8 - 16
  • [30] A Locality Aware Convolutional Neural Networks Accelerator
    Shi, Runbin
    Xu, Zheng
    Sun, Zhihao
    Wu, Di
    Peemen, Maurice
    Li, Ang
    Corporaal, Henk
    2015 EUROMICRO CONFERENCE ON DIGITAL SYSTEM DESIGN (DSD), 2015, : 591 - 598