Hardware Efficient Convolution Processing Unit for Deep Neural Networks

被引:1
|
作者
Hazarika, Anakhi [1 ]
Poddar, Soumyajit [1 ]
Rahaman, Hafizur [2 ]
机构
[1] Indian Inst Informat Technol Guwahati, Gauhati 781015, India
[2] Indian Inst Engn Sci & Technol, Sibpur 711103, Howrah, India
关键词
Deep Neural Network; CNN Hardware Accelerator; Field Programmable Gate Array (FPGA);
D O I
10.1109/isdcs.2019.8719278
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Convolutional Neural Network (CNN) is a type of deep neural networks that are commonly used for object detection and classification. State-of-the-art hardware for training and inference of CNN architectures require a considerable amount of computation and memory intensive resources. CNN achieves greater accuracy at the cost of high computational complexity and large power consumption. To optimize the memory requirement, processing speed and power, it is crucial to design more efficient accelerator architecture for CNN computation. In this work, an overlap of spatially adjacent data is exploited in order to parallelize the movement of data. A fast, re-configurable hardware accelerator architecture along with optimized kernel design suitable for a variety of CNN models is proposed. Our design achieves 2.1x computational benefits over state-of-the-art accelerator architectures.
引用
收藏
页数:4
相关论文
共 50 条
  • [1] Architecture of neural processing unit for deep neural networks
    Lee, Kyuho J.
    HARDWARE ACCELERATOR SYSTEMS FOR ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING, 2021, 122 : 217 - 245
  • [2] Efficient Softmax Hardware Architecture for Deep Neural Networks
    Du, Gaoming
    Tian, Chao
    Li, Zhenmin
    Zhang, Duoli
    Yin, Yongsheng
    Ouyang, Yiming
    GLSVLSI '19 - PROCEEDINGS OF THE 2019 ON GREAT LAKES SYMPOSIUM ON VLSI, 2019, : 75 - 80
  • [3] TermiNETor: Early Convolution Termination for Efficient Deep Neural Networks
    Mallappa, Uday
    Gangwar, Pranav
    Khaleghi, Behnam
    Yang, Haichao
    Rosing, Tajana
    2022 IEEE 40TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD 2022), 2022, : 635 - 643
  • [4] Winograd Convolution for Deep Neural Networks: Efficient Point Selection
    Alam, Syed Asad
    Anderson, Andrew
    Barabasz, Barbara
    Gregg, David
    ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2022, 21 (06)
  • [5] Splatter: An Efficient Sparse Image Convolution for Deep Neural Networks
    Lee, Tristan
    Lee, Byeong Kil
    2021 INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND COMPUTATIONAL INTELLIGENCE (CSCI 2021), 2021, : 506 - 509
  • [6] Efficient Hardware Architectures for Accelerating Deep Neural Networks: Survey
    Dhilleswararao, Pudi
    Boppu, Srinivas
    Manikandan, M. Sabarimalai
    Cenkeramaddi, Linga Reddy
    IEEE ACCESS, 2022, 10 : 131788 - 131828
  • [7] Efficient Processing of Deep Neural Networks: A Tutorial and Survey
    Sze, Vivienne
    Chen, Yu-Hsin
    Yang, Tien-Ju
    Emer, Joel S.
    PROCEEDINGS OF THE IEEE, 2017, 105 (12) : 2295 - 2329
  • [8] An Energy-efficient Convolution Unit for Depthwise Separable Convolutional Neural Networks
    Chong, Yi Sheng
    Goh, Wang Ling
    Ong, Yew Soon
    Nambiar, Vishnu P.
    Do, Anh Tuan
    2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2021,
  • [9] Algorithms and Hardware for Efficient Processing of Logic-based Neural Networks
    Hong, Jingkai
    Fayyazi, Arash
    Esmaili, Amirhossein
    Nazemi, Mandi
    Pedram, Massoud
    2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC, 2023,
  • [10] Hybrid Convolution Architecture for Energy-Efficient Deep Neural Network Processing
    Kim, Suchang
    Jo, Jihyuck
    Park, In-Cheol
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2021, 68 (05) : 2017 - 2029