Design of high performance and energy efficient convolution array for convolution neural network-based image inference engine

被引:1
|
作者
Deepika, S. [1 ]
Arunachalam, V. [1 ]
机构
[1] Vellore Inst Technol, Dept Micro & Nanoelect, Vellore 632014, India
关键词
Convolutional neural network; Edge computing; Energy efficient convolution array; Inference engine; Sparse accelerator; ACCELERATOR;
D O I
10.1016/j.engappai.2023.106953
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The energy efficiency of CNN-based inference engines predominately depends upon Giga-operations-per-second and power consumption. The sparse-based accelerator compresses the insignificant inputs (input feature maps & weights), skips the inefficient computations, and improves energy efficiency. A sparse accelerator for weights could impact the accuracy of the inference. Therefore, a sparse network for Input Feature Maps (IFMs) is considered. MATLAB-based sparsity analysis is done layer-wise on the pre-trained CNN models like AlexNet, VGG-16, VGG-19, ResNet-18 & ResNet-34. Layer-wise analysis reveals that similar to 18%-90% of the IFMs are zeros. Besides, IFMs and Weights adopted the 16-bit Fix/Float data format to maintain an accuracy as close as 97% with Single Precision Floating Point (SPFP). A 3 x 1 Convolutional array with improved Zero-detect-Skip (CZS(3x1)) control units for multiplier and adder/subtractor arrays is proposed. The modified rectified linear unit (RELU) converts IFM values <= 0 to zero and sets Detection-Bit (DB) to 1. These DBs decide the mode of effective zero-skip operations in CZS(3x1). A 3 x 3 Compressed Processing Element (CPE) is designed using three CZS(3x1). The 20-CPEs convolution array architecture is implemented in 65 nm technology libraries. The performance of 90 Giga-operations per second (GOP/s) and energy efficiency of 3.42 Tera operations per second per watt (TOPS/W) were attained for the proposed CPE. The CPE with improved control strategy enhanced the performance by a factor of 2.45 while consuming 8.8 times less energy on average than the state-of-the-art CNN accelerators.
引用
收藏
页数:13
相关论文
共 50 条
  • [41] Deep Convolution Neural Network-Based Crack Feature Extraction, Detection and Quantification
    Teng, Shuai
    Chen, Gongfa
    JOURNAL OF FAILURE ANALYSIS AND PREVENTION, 2022, 22 (03) : 1308 - 1321
  • [42] Wood Microscopic Image Identification Method Based on Convolution Neural Network
    Zhao, Ziyu
    Yang, Xiaoxia
    Ge, Zhedong
    Guo, Hui
    Zhou, Yucheng
    BIORESOURCES, 2021, 16 (03): : 4986 - 4999
  • [43] Secret Image Sharing Schemes Based on Region Convolution Neural Network
    Liu Y.
    Wu P.
    Sun Q.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2021, 58 (05): : 1065 - 1074
  • [44] Research on Classification of Architectural Style Image Based on Convolution Neural Network
    Guo, Kun
    Li, Ning
    2017 IEEE 3RD INFORMATION TECHNOLOGY AND MECHATRONICS ENGINEERING CONFERENCE (ITOEC), 2017, : 1062 - 1066
  • [45] Research on image classification model based on deep convolution neural network
    Xin, Mingyuan
    Wang, Yong
    EURASIP JOURNAL ON IMAGE AND VIDEO PROCESSING, 2019, 2019 (1)
  • [46] Hyperspectral Image Classification based on Spectral Deformable Convolution Neural Network
    Xue Z.
    Li B.
    National Remote Sensing Bulletin, 2022, 26 (10) : 2014 - 2028
  • [47] Hyperspectral Image Classification Based on Convolution Neural Network with Attention Mechanism
    Chen Wenhao
    Jing, He
    Gang, Liu
    LASER & OPTOELECTRONICS PROGRESS, 2022, 59 (18)
  • [48] Satellite Image Matching Method Based on Deep Convolution Neural Network
    Fan D.
    Dong Y.
    Zhang Y.
    2018, SinoMaps Press (47): : 844 - 853
  • [49] Sonar Image Target Detection and Recognition Based on Convolution Neural Network
    Wu Yanchen
    MOBILE INFORMATION SYSTEMS, 2021, 2021
  • [50] Banknote Image Defect Recognition Method Based on Convolution Neural Network
    Wang Ke
    Wang Huiqin
    Shu Yue
    Mao Li
    Qiu Fengyan
    INTERNATIONAL JOURNAL OF SECURITY AND ITS APPLICATIONS, 2016, 10 (06): : 269 - 279