An FPGA Realization of a Deep Convolutional Neural Network Using a Threshold Neuron Pruning

被引:6
|
作者
Fujii, Tomoya [1 ]
Sato, Simpei [1 ]
Nakahara, Hiroki [1 ]
Motomura, Masato [2 ]
机构
[1] Tokyo Inst Technol, Meguro Ku, Tokyo, Japan
[2] Hokkaido Univ, Sapporo, Hokkaido, Japan
来源
关键词
D O I
10.1007/978-3-319-56258-2_23
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
For a pre-trained deep convolutional neural network (CNN) for an embedded system, a high-speed and a low power consumption are required. In the former of the CNN, it consists of convolutional layers, while in the latter, it consists of fully connection layers. In the convolutional layer, the multiply accumulation operation is a bottleneck, while the fully connection layer, the memory access is a bottleneck. In this paper, we propose a neuron pruning technique which eliminates almost part of the weight memory. In that case, the weight memory is realized by an on-chip memory on the FPGA. Thus, it achieves a high speed memory access. In this paper, we propose a sequential-input parallel-output fully connection layer circuit. The experimental results showed that, by the neuron pruning, as for the fully connected layer on the VGG-11 CNN, the number of neurons was reduced by 89.3% with keeping the 99% accuracy. We implemented the fully connected layers on the Digilent Inc. NetFPGA-1G-CML board. Comparison with the CPU (ARM Cortex A15 processor) and the GPU (Jetson TK1 Kepler), as for a delay time, the FPGA was 219.0 times faster than the CPU and 12.5 times faster than the GPU. Also, a performance per power efficiency was 125.28 times better than CPU and 17.88 times better than GPU.
引用
收藏
页码:268 / 280
页数:13
相关论文
共 50 条
  • [21] A Z Structure Convolutional Neural Network Implemented by FPGA in Deep Learning
    Zhu, Min
    Kuang, Qiqi
    Lin, JianJun
    Luo, Qihong
    Yang, Chunling
    Liu, Ming
    [J]. IECON 2018 - 44TH ANNUAL CONFERENCE OF THE IEEE INDUSTRIAL ELECTRONICS SOCIETY, 2018, : 2677 - 2682
  • [22] FPGA Based Reconfigurable Coprocessor for Deep Convolutional Neural Network Training
    Clere, Sajna Remi
    Sachin, S.
    Varghese, Kuruvilla
    [J]. 2018 21ST EUROMICRO CONFERENCE ON DIGITAL SYSTEM DESIGN (DSD 2018), 2018, : 381 - 388
  • [23] A Fast FPGA-based Deep Convolutional Neural Network Using Pseudo Parallel Memories
    Hailesellasie, Muluken
    Hasan, Syed Rafay
    [J]. 2017 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2017, : 364 - 367
  • [24] FPGA implementation of epileptic seizure detection using semisupervised reduced deep convolutional neural network
    Sahani, Mrutyunjaya
    Rout, Susanta Kumar
    Dash, Pradipta Kishore
    [J]. APPLIED SOFT COMPUTING, 2021, 110
  • [25] Efficient multiquality super-resolution using a deep convolutional neural network for an FPGA implementation
    Kim, Min Beom
    Lee, Sanglyn
    Kim, Ilho
    Hong, Hee Jung
    Kim, Chang Gone
    Yoon, Soo Young
    [J]. JOURNAL OF THE SOCIETY FOR INFORMATION DISPLAY, 2020, 28 (05) : 428 - 439
  • [26] An Efficient FPGA-based Depthwise Separable Convolutional Neural Network Accelerator with Hardware Pruning
    Liu, Zhengyan
    Liu, Qiang
    Yan, Shun
    Cheung, Ray C. C.
    [J]. ACM TRANSACTIONS ON RECONFIGURABLE TECHNOLOGY AND SYSTEMS, 2024, 17 (01)
  • [27] Pruning Convolutional Neural Network with Distinctiveness Approach
    Li, Wenrui
    Plested, Jo
    [J]. NEURAL INFORMATION PROCESSING, ICONIP 2019, PT V, 2019, 1143 : 448 - 455
  • [28] Rethinking the Pruning Criteria for Convolutional Neural Network
    Huang, Zhongzhan
    Shao, Wenqi
    Wang, Xinjiang
    Lin, Liang
    Luo, Ping
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [29] Thinning of convolutional neural network with mixed pruning
    Yang, Wenzhu
    Jin, Lilei
    Wang, Sile
    Cu, Zhenchao
    Chen, Xiangyang
    Chen, Liping
    [J]. IET IMAGE PROCESSING, 2019, 13 (05) : 779 - 784
  • [30] Compression of Deep Convolutional Neural Network Using Additional Importance-Weight-Based Filter Pruning Approach
    Sawant, Shrutika S.
    Wiedmann, Marco
    Goeb, Stephan
    Holzer, Nina
    Lang, Elmar W.
    Goetz, Theresa
    [J]. APPLIED SCIENCES-BASEL, 2022, 12 (21):