A Pipelined and Scalable Dataflow Implementation of Convolutional Neural Networks on FPGA

被引:11
|
作者
Bacis, Marco [1 ]
Natale, Giuseppe [1 ]
Del Sozzo, Emanuele [1 ]
Santambrogio, Marco Domenico [1 ]
机构
[1] Politecn Milan, Dipartimento Elettron Informaz & Bioingn, Milan, Italy
关键词
Field Programmable Gate Arrays; Convolutional Neural Networks; Dataflow Architectures; COPROCESSOR; PERFORMANCE;
D O I
10.1109/IPDPSW.2017.44
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Convolutional Neural Network (CNN) is a deep learning algorithm extended from Artificial Neural Network (ANN) and widely used for image classification and recognition, thanks to its invariance to distortions. The recent rapid growth of applications based on deep learning algorithms, especially in the context of Big Data analytics, has dramatically improved both industrial and academic research and exploration of optimized implementations of CNNs on accelerators such as GPUs, FPGAs and ASICs, as general purpose processors can hardly meet the ever increasing performance and energy-efficiency requirements. FPGAs in particular are one of the most attractive alternative, as they allow the exploitation of the implicit parallelism of the algorithm and the acceleration of the different layers of a CNN with custom optimizations, while retaining extreme flexibility thanks to their reconfigurability. In this work, we propose a methodology to implement CNNs on FPGAs in a modular, scalable way. This is done by exploiting the dataflow pattern of convolutions, using an approach derived from previous work on the acceleration of Iterative Stencil Loops (ISLs), a computational pattern that shares some characteristics with convolutions. Furthermore, this approach allows the implementation of a high-level pipeline between the different network layers, resulting in an increase of the overall performance when the CNN is employed to process batches of multiple images, as it would happen in real-life scenarios.
引用
下载
收藏
页码:90 / 97
页数:8
相关论文
共 50 条
  • [31] FPGA based Flexible Implementation of Light Weight Inference on Deep Convolutional Neural Networks
    Dawwd, Shefa
    INTERNATIONAL ARAB JOURNAL OF INFORMATION TECHNOLOGY, 2024, 21 (03) : 408 - 417
  • [32] Application of Bit-Serial Arithmetic Units for FPGA Implementation of Convolutional Neural Networks
    Csordas, G.
    Feher, B.
    Kovacshazy, T.
    2018 19TH INTERNATIONAL CARPATHIAN CONTROL CONFERENCE (ICCC), 2018, : 322 - 327
  • [33] Abstraction in FPGA implementation of neural networks
    Ogrenci, Arif Selcuk
    PROCEEDINGS OF THE 9TH WSEAS INTERNATIONAL CONFERENCE ON NEURAL NETWORKS (NN' 08): ADVANCED TOPICS ON NEURAL NETWORKS, 2008, : 221 - 224
  • [34] Reconfigurable FPGA implementation of neural networks
    Hajduk, Zbigniew
    NEUROCOMPUTING, 2018, 308 : 227 - 234
  • [35] Pipelined Training with Stale Weights in Deep Convolutional Neural Networks
    Zhang, Lifu
    Abdelrahman, Tarek S.
    APPLIED COMPUTATIONAL INTELLIGENCE AND SOFT COMPUTING, 2021, 2021
  • [36] FPGA-Implementation of Pipelined Neural Network for Power Amplifier Modeling
    Ntoune, Roger Sandrin Ntoune
    Bahoura, Mohammed
    Park, Chan-Wang
    2012 IEEE 10TH INTERNATIONAL NEW CIRCUITS AND SYSTEMS CONFERENCE (NEWCAS), 2012, : 109 - 112
  • [37] Acceleration and Implementation of Convolutional Neural Network Based on FPGA
    Wang, Enyi
    Qiu, Dehui
    PROCEEDINGS OF 2019 IEEE 7TH INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND NETWORK TECHNOLOGY (ICCSNT 2019), 2019, : 321 - 325
  • [38] Design and Implementation of Configurable Convolutional Neural Network on FPGA
    Huynh Vinh Phu
    Tran Minh Tan
    Phan Van Men
    Nguyen Van Hieu
    Truong Van Cuong
    PROCEEDINGS OF 2019 6TH NATIONAL FOUNDATION FOR SCIENCE AND TECHNOLOGY DEVELOPMENT (NAFOSTED) CONFERENCE ON INFORMATION AND COMPUTER SCIENCE (NICS), 2019, : 298 - 302
  • [39] Evaluation of Dataflow through layers of convolutional neural networks in classification problems
    Saffar, Mohsen
    Kalhor, Ahmad
    EXPERT SYSTEMS WITH APPLICATIONS, 2023, 224
  • [40] SparseTrain: Exploiting Dataflow Sparsity for Efficient Convolutional Neural Networks Training
    Dai, Pengcheng
    Yang, Jianlei
    Ye, Xucheng
    Cheng, Xingzhou
    Luo, Junyu
    Song, Linghao
    Chen, Yiran
    Zhao, Weisheng
    PROCEEDINGS OF THE 2020 57TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2020,