Efficient Hardware Acceleration of Convolutional Neural Networks

被引:0
|
作者
Kala, S. [1 ]
Jose, Babita R. [1 ]
Mathew, Jimson [2 ]
Nalesh, S. [3 ]
机构
[1] Cochin Univ Sci & Technol, Sch Engn, Kochi, Kerala, India
[2] Indian Inst Technol Patna, Dept Comp Sci & Engn, Patna, Bihar, India
[3] Cochin Univ Sci & Technol, Dept Elect, Cochin, Kerala, India
关键词
Convolutional neural networks; FPGA; high performance; Winograd algorithm;
D O I
10.1109/SOCC46988.2019.1570573948
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Convolutional neural networks (CNNs) have emerged as the most efficient technique for solving a host of machine learning tasks. Compute and memory intensive nature of CNN has stimulated lot of work in hardware acceleration of these network models. FPGAs have emerged as a promising approach for accelerating CNNs, due to its high performance, flexibility and energy efficiency. We propose a unified architecture named UniWiG, where both Winograd based convolution and general matrix multiplication (GEMM) can be accelerated using the same set of processing elements. Proposed architecture has been used to accelerate AlexNet and VGG-16 models on FPGA with a performance of 433.63 GOPS and 407.23 GOPS respectively. We have also analyzed the performance with varying Winograd tile sizes and found out the most appropriate tile sizes for maximizing the performance while reducing on-chip memory resource.
引用
收藏
页码:191 / 192
页数:2
相关论文
共 50 条
  • [21] Hardware Acceleration of Graph Neural Networks
    Auten, Adam
    Tomei, Matthew
    Kumar, Rakesh
    [J]. PROCEEDINGS OF THE 2020 57TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2020,
  • [22] HFP: Hardware-Aware Filter Pruning for Deep Convolutional Neural Networks Acceleration
    Yu, Fang
    Han, Chuanqi
    Wang, Pengcheng
    Huang, Ruoran
    Huang, Xi
    Cui, Li
    [J]. 2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 255 - 262
  • [23] HILP: hardware-in-loop pruning of convolutional neural networks towards inference acceleration
    Dong Li
    Qianqian Ye
    Xiaoyue Guo
    Yunda Sun
    Li Zhang
    [J]. Neural Computing and Applications, 2024, 36 : 8825 - 8842
  • [24] HILP: hardware-in-loop pruning of convolutional neural networks towards inference acceleration
    Li, Dong
    Ye, Qianqian
    Guo, Xiaoyue
    Sun, Yunda
    Zhang, Li
    [J]. NEURAL COMPUTING & APPLICATIONS, 2024, 36 (15): : 8825 - 8842
  • [25] A Pipelined Energy-efficient Hardware Accelaration for Deep Convolutional Neural Networks
    Alaeddine, Hmidi
    Jihene, Malek
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON DESIGN & TEST OF INTEGRATED MICRO & NANO-SYSTEMS (DTS), 2019,
  • [26] An Updated Survey of Efficient Hardware Architectures for Accelerating Deep Convolutional Neural Networks
    Capra, Maurizio
    Bussolino, Beatrice
    Marchisio, Alberto
    Shafique, Muhammad
    Masera, Guido
    Martina, Maurizio
    [J]. FUTURE INTERNET, 2020, 12 (07):
  • [27] Energy-efficient acceleration of convolutional neural networks using computation reuse
    Ghanbari, Azam
    Modarressi, Mehdi
    [J]. JOURNAL OF SYSTEMS ARCHITECTURE, 2022, 126
  • [28] ASBNN: Acceleration of Bayesian Convolutional Neural Networks by Algorithm-hardware Co-design
    Fujiwara, Yoshiki
    Takamaeda-Yamazaki, Shinya
    [J]. 2021 IEEE 32ND INTERNATIONAL CONFERENCE ON APPLICATION-SPECIFIC SYSTEMS, ARCHITECTURES AND PROCESSORS (ASAP 2021), 2021, : 226 - 233
  • [29] VIBNN: Hardware Acceleration of Bayesian Neural Networks
    Cai, Ruizhe
    Ren, Ao
    Liu, Ning
    Ding, Caiwen
    Wang, Luhao
    Qian, Xuehai
    Pedram, Massoud
    Wang, Yanzhi
    [J]. ACM SIGPLAN NOTICES, 2018, 53 (02) : 476 - 488
  • [30] Hardware Design Automation of Convolutional Neural Networks
    Solazzo, Andrea
    Del Sozzo, Emanuele
    De Rose, Irene
    De Silvestri, Matteo
    Durelli, Gianluca C.
    Santambrogio, Marco D.
    [J]. 2016 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI), 2016, : 224 - 229