SaARSP: An Architecture for Systolic-Array Acceleration of Recurrent Spiking Neural Networks

被引:3
|
作者
Lee, Jeong-Jun [1 ]
Zhang, Wenrui [1 ]
Xie, Yuan [1 ]
Li, Peng [1 ]
机构
[1] Univ Calif Santa Barbara, Elect & Comp Engn, Santa Barbara, CA 93106 USA
基金
美国国家科学基金会;
关键词
Spiking neural networks; accelerators; computer architecture; IMPLEMENTATION; NEURONS; FLOW;
D O I
10.1145/3510854
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Spiking neural networks (SNNs) are brain-inspired event-driven models of computation with promising ultra-low energy dissipation. Rich network dynamics emergent in recurrent spiking neural networks (R-SNNs) can form temporally based memory, offering great potential in processing complex spatiotemporal data. However, recurrence in network connectivity produces tightly coupled data dependency in both space and time, rendering hardware acceleration of R-SNNs challenging. We present the first work to exploit spatiotemporal parallelisms to accelerate the R-SNN-based inference on systolic arrays using an architecture called SaARSP. We decouple the processing of feedforward synaptic connections from that of recurrent connections to allow for the exploitation of parallelisms across multiple time points. We propose a novel time window size optimization (TWSO) technique, to further explore the temporal granularity of the proposed decoupling in terms of optimal time window size and reconfiguration of the systolic array considering layer-dependent connectivity to boost performance. Stationary dataflow and time window size are jointly optimized to trade off between weight data reuse and movements of partial sums, the two bottlenecks in latency and energy dissipation of the accelerator. The proposed systolic-array architecture offers a unifying solution to an acceleration of both feedforward and recurrent SNNs, and delivers 4,000X EDP improvement on average for different R-SNN benchmarks over a conventional baseline.
引用
收藏
页数:23
相关论文
共 50 条
  • [1] Parallel Time Batching: Systolic-Array Acceleration of Sparse Spiking Neural Computation
    Lee, Jeong-Jun
    Zhang, Wenrui
    Li, Peng
    [J]. 2022 IEEE INTERNATIONAL SYMPOSIUM ON HIGH-PERFORMANCE COMPUTER ARCHITECTURE (HPCA 2022), 2022, : 317 - 330
  • [2] Systolic-Array Spiking Neural Accelerators with Dynamic Heterogeneous Voltage Regulation
    Lee, Jeong-Jun
    Chen, Jianhao
    Zhang, Wenrui
    Li, Peng
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [3] AN ADVANCED DSP SYSTOLIC-ARRAY ARCHITECTURE
    LEELAND, SB
    [J]. COMPUTER, 1987, 20 (07) : 95 - 96
  • [4] An instruction systolic array architecture for neural networks
    Kane, AJ
    Evans, DJ
    [J]. INTERNATIONAL JOURNAL OF COMPUTER MATHEMATICS, 1996, 61 (1-2) : 63 - 89
  • [5] Efficient Systolic-Array Redundancy Architecture for Offline/Online Repair
    Cho, Keewon
    Lee, Ingeol
    Lim, Hyeonchan
    Kang, Sungho
    [J]. ELECTRONICS, 2020, 9 (02)
  • [6] FIBEROPTIC MATRIX MULTIPLIER USING A TWO-DIMENSIONAL SYSTOLIC-ARRAY ARCHITECTURE
    SHABEER, M
    ANDONOVIC, I
    CULSHAW, B
    [J]. OPTICS LETTERS, 1987, 12 (11) : 959 - 961
  • [7] Neural Architecture Search for Spiking Neural Networks
    Kim, Youngeun
    Li, Yuhang
    Park, Hyoungseob
    Venkatesha, Yeshwanth
    Panda, Priyadarshini
    [J]. COMPUTER VISION, ECCV 2022, PT XXIV, 2022, 13684 : 36 - 56
  • [8] A Novel Systolic Parallel Hardware Architecture for the FPGA Acceleration of Feedforward Neural Networks
    Medus, Leandro D.
    Iakymchuk, Taras
    Frances-Villora, Jose Vicente
    Bataller-Mompean, Manuel
    Rosado-Munoz, Alfredo
    [J]. IEEE ACCESS, 2019, 7 : 76084 - 76103
  • [9] Two-Dimensional Systolic-Array Architecture for Pixel-Level Vision Tasks
    Vijverberg, Julien A.
    de With, Peter H. N.
    [J]. REAL-TIME IMAGE AND VIDEO PROCESSING 2010, 2010, 7724
  • [10] Configurable Multi-directional Systolic Array Architecture for Convolutional Neural Networks
    Xu, Rui
    Ma, Sheng
    Wang, Yaohua
    Chen, Xinhai
    Guo, Yang
    [J]. ACM TRANSACTIONS ON ARCHITECTURE AND CODE OPTIMIZATION, 2021, 18 (04)