FiC-RNN: A Multi-FPGA Acceleration Framework for Deep Recurrent Neural Networks

被引:12
|
作者
Sun, Yuxi [1 ]
Amano, Hideharu [1 ]
机构
[1] Keio Univ, Dept Informat & Comp Sci, Yokohama, Kanagawa 2238522, Japan
关键词
multi-FPGA; recurrent neural networks; LSTM;
D O I
10.1587/transinf.2020PAP0003
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Recurrent neural networks (RNNs) have been proven effective for sequence-based tasks thanks to their capability to process temporal information. In real-world systems, deep RNNs are more widely used to solve complicated tasks such as large-scale speech recognition and machine translation. However, the implementation of deep RNNs on traditional hardware platforms is inefficient due to long-range temporal dependence and irregular computation patterns within RNNs. This inefficiency manifests itself in the proportional increase in the latency of RNN inference with respect to the number of layers of deep RNNs on CPUs and GPUs. Previous work has focused mostly on optimizing and accelerating individual RNN cells. To make deep RNN inference fast and efficient, we propose an accelerator based on a multi-FPGA platform called Flow-inCloud (FiC). In this work, we show that the parallelism provided by the multi-FPGA system can be taken advantage of to scale up the inference of deep RNNs, by partitioning a large model onto several FPGAs, so that the latency stays close to constant with respect to increasing number of RNN layers. For single-layer and four-layer RNNs, our implementation achieves 31x and 61x speedup compared with an Intel CPU.
引用
收藏
页码:2457 / 2462
页数:6
相关论文
共 50 条
  • [41] TS-RNN: Text Steganalysis Based on Recurrent Neural Networks
    Yang, Zhongliang
    Wang, Ke
    Li, Jian
    Huang, Yongfeng
    Zhang, Yu-Jin
    IEEE SIGNAL PROCESSING LETTERS, 2019, 26 (12) : 1743 - 1747
  • [42] A Deep Learning Framework on Generation of Image Descriptions with Bidirectional Recurrent Neural Networks
    Thomas, J. Joshua
    Pillai, Naris
    INTELLIGENT COMPUTING & OPTIMIZATION, 2019, 866 : 219 - 230
  • [43] A 'programming' framework for recurrent neural networks
    Beiran, Manuel
    Spencer-Salmon, Camille A.
    Rajan, Kanaka
    NATURE MACHINE INTELLIGENCE, 2023, 5 (6) : 570 - 571
  • [44] A ‘programming’ framework for recurrent neural networks
    Manuel Beiran
    Camille A. Spencer-Salmon
    Kanaka Rajan
    Nature Machine Intelligence, 2023, 5 : 570 - 571
  • [45] A Unified FPGA Virtualization Framework for General-Purpose Deep Neural Networks in the Cloud
    Zeng, Shulin
    Dai, Guohao
    Sun, Hanbo
    Liu, Jun
    Li, Shiyao
    Ge, Guangjun
    Zhong, Kai
    Guo, Kaiyuan
    Wang, Yu
    Yang, Huazhong
    ACM TRANSACTIONS ON RECONFIGURABLE TECHNOLOGY AND SYSTEMS, 2022, 15 (03)
  • [46] An FSCV Deep Neural Network: Development, Pruning, and Acceleration on an FPGA
    Zhang, Zhichao
    Oh, Yoonbae
    Adams, Scott D.
    Bennet, Kevin E.
    Kouzani, Abbas Z.
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2021, 25 (06) : 2248 - 2259
  • [47] A High-level Implementation Framework for Non-Recurrent Artificial Neural Networks on FPGA
    Prono, Luciano
    Marchioni, Alex
    Mangia, Mauro
    Pareschi, Fabio
    Rovatti, Riccardo
    Setti, Gianluca
    2019 15TH CONFERENCE ON PHD RESEARCH IN MICROELECTRONICS AND ELECTRONICS (PRIME), 2019, : 77 - 80
  • [48] Continuously Deep Recurrent Neural Networks
    Ceni, Andrea
    Dominey, Peter Ford
    Gallicchio, Claudio
    Micheli, Alessio
    Pedrelli, Luca
    Tortorella, Domenico
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, PT VII, ECML PKDD 2024, 2024, 14947 : 59 - 73
  • [49] Recurrent diffractive deep neural networks
    Zhou, Junhe
    Wang, Qiqi
    Huang, Chenweng
    OPTICS EXPRESS, 2024, 32 (27): : 48093 - 48104
  • [50] Accelerating Deep Neural Networks Using FPGA
    Adel, Esraa
    Magdy, Rana
    Mohamed, Sara
    Mamdouh, Mona
    El Mandouh, Eman
    Mostafa, Hassan
    2018 30TH INTERNATIONAL CONFERENCE ON MICROELECTRONICS (ICM), 2018, : 176 - 179