Efficient Hardware Acceleration of Convolutional Neural Networks

被引:0
|
作者
Kala, S. [1 ]
Jose, Babita R. [1 ]
Mathew, Jimson [2 ]
Nalesh, S. [3 ]
机构
[1] Cochin Univ Sci & Technol, Sch Engn, Kochi, Kerala, India
[2] Indian Inst Technol Patna, Dept Comp Sci & Engn, Patna, Bihar, India
[3] Cochin Univ Sci & Technol, Dept Elect, Cochin, Kerala, India
关键词
Convolutional neural networks; FPGA; high performance; Winograd algorithm;
D O I
10.1109/SOCC46988.2019.1570573948
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Convolutional neural networks (CNNs) have emerged as the most efficient technique for solving a host of machine learning tasks. Compute and memory intensive nature of CNN has stimulated lot of work in hardware acceleration of these network models. FPGAs have emerged as a promising approach for accelerating CNNs, due to its high performance, flexibility and energy efficiency. We propose a unified architecture named UniWiG, where both Winograd based convolution and general matrix multiplication (GEMM) can be accelerated using the same set of processing elements. Proposed architecture has been used to accelerate AlexNet and VGG-16 models on FPGA with a performance of 433.63 GOPS and 407.23 GOPS respectively. We have also analyzed the performance with varying Winograd tile sizes and found out the most appropriate tile sizes for maximizing the performance while reducing on-chip memory resource.
引用
收藏
页码:191 / 192
页数:2
相关论文
共 50 条
  • [41] TileNET: Hardware accelerator for ternary Convolutional Neural Networks
    Eetha, Sagar
    Sruthi, P. K.
    Pant, Vibha
    Vikram, Sai
    Mody, Mihir
    Purnaprajna, Madhura
    [J]. MICROPROCESSORS AND MICROSYSTEMS, 2021, 83
  • [42] A Fourier domain acceleration framework for convolutional neural networks
    Lin, Jinhua
    Ma, Lin
    Yao, Yu
    [J]. NEUROCOMPUTING, 2019, 364 : 254 - 268
  • [43] Acceleration and implementation of convolutional neural networks based on FPGA
    Zhao, Sijie
    Gao, Shangshang
    Wang, Rugang
    Wang, Yuanyuan
    Zhou, Feng
    Guo, Naihong
    [J]. DIGITAL SIGNAL PROCESSING, 2023, 141
  • [44] Efficient Hardware Realization of Convolutional Neural Networks using Intra-Kernel Regular Pruning
    Yang, Maurice
    Faraj, Mahmoud
    Hussein, Assem
    Gaudet, Vincent
    [J]. 2018 IEEE 48TH INTERNATIONAL SYMPOSIUM ON MULTIPLE-VALUED LOGIC (ISMVL 2018), 2018, : 180 - 185
  • [45] SOFTWARE-HARDWARE CODESIGN FOR EFFICIENT NEURAL NETWORK ACCELERATION
    Guo, Kaiyuan
    Han, Song
    Yao, Song
    Wang, Yu
    Xie, Yuan
    Yang, Huazhong
    [J]. IEEE MICRO, 2017, 37 (02) : 18 - 25
  • [46] Efficient Hardware Implementation of Threshold Neural Networks
    Zamanlooy, Babak
    Mirhassani, Mitra
    [J]. 2012 IEEE 10TH INTERNATIONAL NEW CIRCUITS AND SYSTEMS CONFERENCE (NEWCAS), 2012, : 1 - 4
  • [47] VWA: Hardware Efficient Vectorwise Accelerator for Convolutional Neural Network
    Chang, Kuo-Wei
    Chang, Tian-Sheuan
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2020, 67 (01) : 145 - 154
  • [48] An Efficient Hardware Volume Renderer for Convolutional Neural Radiance Fields
    Wang, Xuexin
    He, Yunxiang
    Zhang, Xiangyu
    Zhou, Pingqiang
    Lou, Xin
    [J]. 2024 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, ISCAS 2024, 2024,
  • [49] Pre-Defined Sparse Neural Networks With Hardware Acceleration
    Dey, Sourya
    Huang, Kuan-Wen
    Beerel, Peter A.
    Chugg, Keith M.
    [J]. IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2019, 9 (02) : 332 - 345
  • [50] Model Compression and Hardware Acceleration for Neural Networks: A Comprehensive Survey
    Deng, Lei
    Li, Guoqi
    Han, Song
    Shi, Luping
    Xie, Yuan
    [J]. PROCEEDINGS OF THE IEEE, 2020, 108 (04) : 485 - 532