FPGA based Flexible Implementation of Light Weight Inference on Deep Convolutional Neural Networks

被引:0
|
作者
Dawwd, Shefa [1 ]
机构
[1] Univ Mosul, Dept Comp Engn, Mosul, Iraq
关键词
Standard convolution; depthwise separable convolution; inference; deep convolutional neural networks; FPGA;
D O I
10.34028/iajit/21/3/5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Standard Convolution (StdConv) is the main technique used in the state of the art Deep Convolutional Neural Networks (DCNNs). Fewer computations are achieved if Depthwise Separable Convolution technique (SepConv) is used as an alternative. A crucial issue in many applications like smart cameras and autonomous vehicles where low latency is essential stems from deploying a lightweight and low cost inference models. An acceptable accuracy should be kept with tolerable computations and memory access load. A flexible architecture for different DCNN convolution types and models is proposed. The flexibility comes from the sharing of one memory access unit with different types of layers regardless of the selected kernel size, by multiplying each weight vector by local operators with variant aperture. Moreover, one depthwise computation unit can be used for both standard and pointwise layers. The learnable parameters are quantized to 8 -bits fixed point representation and that gives very limited reduction of accuracy and a considerable reduction of the Field -Programmable Gate Array (FPGA) resources. To reduce processing time, inter layer parallel computations are performed. The experiment is conducted by using grey scale ORL database with shallow Convolutional Neural Network (CNN) and the colored Canadian Institute for Advanced Research 10 classes (CIFAR-10) database with DCNN, and a comparable accuracies of 93% and 85.7% are achieved respectively using very low cost of Spartan 3E and moderate cost of zynq FPGA platforms .
引用
收藏
页码:408 / 417
页数:10
相关论文
共 50 条
  • [1] An FPGA-based Accelerator Implementation for Deep Convolutional Neural Networks
    Zhou, Yongmei
    Jiang, Jingfei
    [J]. PROCEEDINGS OF 2015 4TH INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND NETWORK TECHNOLOGY (ICCSNT 2015), 2015, : 829 - 832
  • [2] A Flexible FPGA-Based Inference Architecture for Pruned Deep Neural Networks
    Posewsky, Thorbjoern
    Ziener, Daniel
    [J]. ARCHITECTURE OF COMPUTING SYSTEMS, 2018, 10793 : 311 - 323
  • [3] Acceleration and implementation of convolutional neural networks based on FPGA
    Zhao, Sijie
    Gao, Shangshang
    Wang, Rugang
    Wang, Yuanyuan
    Zhou, Feng
    Guo, Naihong
    [J]. DIGITAL SIGNAL PROCESSING, 2023, 141
  • [4] An Adaptive Row-based Weight Reuse Scheme for FPGA Implementation of Convolutional Neural Networks
    Je, Hyeonseung
    Duy Thanh Nguyen
    Lee, Kyujoong
    Lee, Hyuk-Jae
    [J]. 2021 36TH INTERNATIONAL TECHNICAL CONFERENCE ON CIRCUITS/SYSTEMS, COMPUTERS AND COMMUNICATIONS (ITC-CSCC), 2021,
  • [5] Flexible Bayesian Inference by Weight Transfer for Robust Deep Neural Networks
    Thi Thu Thao Khong
    Nakada, Takashi
    Nakashima, Yasuhiko
    [J]. IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2021, E104D (11) : 1981 - 1991
  • [6] Efficient Implementation of Convolutional Neural Networks on FPGA
    Hadnagy, A.
    Feher, B.
    Kovacshazy, T.
    [J]. 2018 19TH INTERNATIONAL CARPATHIAN CONTROL CONFERENCE (ICCC), 2018, : 359 - 364
  • [7] FPGA Implementation and Acceleration of Convolutional Neural Networks
    Pisharody, Jayanth N.
    Pranav, K. B.
    Ranjitha, M.
    Rajeshwari, B.
    [J]. 2021 6TH INTERNATIONAL CONFERENCE FOR CONVERGENCE IN TECHNOLOGY (I2CT), 2021,
  • [8] Noise Convolutional Neural Networks and FPGA Implementation
    Munakala, Atsuki
    Nakahara, IIiroki
    Sato, Shimpei
    [J]. 2019 IEEE 49TH INTERNATIONAL SYMPOSIUM ON MULTIPLE-VALUED LOGIC (ISMVL), 2019, : 85 - 90
  • [9] xDNN: Inference for Deep Convolutional Neural Networks
    D'Alberto, Paolo
    Wu, Victor
    Ng, Aaron
    Nimaiyar, Rahul
    Delaye, Elliott
    Sirasao, Ashish
    [J]. ACM TRANSACTIONS ON RECONFIGURABLE TECHNOLOGY AND SYSTEMS, 2022, 15 (02)
  • [10] Implementation of energy-efficient fast convolution algorithm for deep convolutional neural networks based on FPGA
    Li, W. -J.
    Ruan, S. -J.
    Yang, D. -S.
    [J]. ELECTRONICS LETTERS, 2020, 56 (10) : 485 - 487