A Fast Compressed Hardware Architecture for Deep Neural Networks

被引:0
|
作者
Ansari, Anaam [1 ]
Shelton, Allen [1 ]
Ogunfunmi, Tokunbo [1 ]
Panchbhaiyye, Vineet [1 ]
机构
[1] Santa Clara Univ, Dept Elect & Comp Engn, Santa Clara, CA 95053 USA
关键词
deep learning; convolutional neural network; hardware architecture; design methodology; edge intelligence; channel pruning; compressed network;
D O I
10.1109/ISCAS48785.2022.9937651
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Hardware acceleration of Deep Neural Networks (DNNs) is very critical to many edge applications. The acceleration solutions available today are typically for GPU, CPU, FPGA and ASIC platforms. The Single Partial Product 2-D Convolution known as SPP2D is a hardware architecture for fast 2-D convolution which can be used for implementing a convolutional neural network (CNN). The SPP2D based CNNs prevent the re-fetching of input pixels for the calculation of partial products and it computes the output for any input size and kernel with low latency and low power consumption compared to some other popular techniques. SPP2D based VGGNet-16 rivals the performance of other existing implementations as well as a FIFO based CNN architecture which takes a novel approach to compute convolution results using row-wise inputs as opposed to traditional tile-based processing. In this paper, we present an SPP2D based hardware accelerator for a channel compressed network. We find that the low power SPP2D implementation gives a better performance in terms of low power and low execution time compared to other compressed contemporary designs. A compressed network requires less on-chip memory thus reducing the most power consuming task of moving data from off-chip to on-chip. This results in a considerable reduction in power consumption due to the reduction in memory traffic. Channelpruned SPP2D accelerator is a low power design of 298 mW which is about 0.01x to 0.37x of the existing works while also having a low execution time of 0.9 secs.
引用
收藏
页码:370 / 374
页数:5
相关论文
共 50 条
  • [41] Fast Convex Pruning of Deep Neural Networks
    Aghasi, Alireza
    Abdi, Afshin
    Romberg, Justin
    [J]. SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2020, 2 (01): : 158 - 188
  • [42] Fast Fingerprint Classification with Deep Neural Networks
    Michelsanti, Daniel
    Ene, Andreea-Daniela
    Guichi, Yanis
    Stef, Rares
    Nasrollahi, Kamal
    Moeslund, Thomas B.
    [J]. PROCEEDINGS OF THE 12TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS (VISIGRAPP 2017), VOL 5, 2017, : 202 - 209
  • [43] RECOM: An Efficient Resistive Accelerator for Compressed Deep Neural Networks
    Ji, Houxiang
    Song, Linghao
    Jiang, Li
    Li, Ha
    Chen, Yiran
    [J]. PROCEEDINGS OF THE 2018 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE), 2018, : 237 - 240
  • [44] Compressed Superposition of Neural Networks for Deep Learning in Edge Computing
    Zeman, Marko
    Osipov, Evgeny
    Bosnic, Zoran
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [45] INVISIBLE AND EFFICIENT BACKDOOR ATTACKS FOR COMPRESSED DEEP NEURAL NETWORKS
    Phan, Huy
    Xie, Yi
    Liu, Jian
    Chen, Yingying
    Yuan, Bo
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 96 - 100
  • [46] A Computing Efficient Hardware Architecture for Sparse Deep Neural Network Computing
    Zhang, Yanwen
    Ouyang, Peng
    Yin, Shouyi
    Zhang, Youguang
    Zhao, Weisheng
    Wei, Shaojun
    [J]. 2018 14TH IEEE INTERNATIONAL CONFERENCE ON SOLID-STATE AND INTEGRATED CIRCUIT TECHNOLOGY (ICSICT), 2018, : 1261 - 1263
  • [47] Data multiplexed and hardware reused architecture for deep neural network accelerator
    Raut, Gopal
    Biasizzo, Anton
    Dhakad, Narendra
    Gupta, Neha
    Papa, Gregor
    Vishvakarma, Santosh Kumar
    [J]. NEUROCOMPUTING, 2022, 486 : 147 - 159
  • [48] Hardware Accelerator for Adversarial Attacks on Deep Learning Neural Networks
    Guo, Haoqiang
    Peng, Lu
    Zhang, Jian
    Qi, Fang
    Duan, Lide
    [J]. 2019 TENTH INTERNATIONAL GREEN AND SUSTAINABLE COMPUTING CONFERENCE (IGSC), 2019,
  • [49] SafeTPU: A Verifiably Secure Hardware Accelerator for Deep Neural Networks
    Mera Collantes, Maria I.
    Ghodsi, Zahra
    Garg, Siddharth
    [J]. Proceedings of the IEEE VLSI Test Symposium, 2020, 2020-April
  • [50] Hardware-Aware Softmax Approximation for Deep Neural Networks
    Geng, Xue
    Lin, Jie
    Zhao, Bin
    Kong, Anmin
    Aly, Mohamed M. Sabry
    Chandrasekhar, Vijay
    [J]. COMPUTER VISION - ACCV 2018, PT IV, 2019, 11364 : 107 - 122