PIE: A Pipeline Energy-efficient Accelerator for Inference Process in Deep Neural Networks

被引:0
|
作者
Zhao, Yangyang [1 ]
Yu, Qi [1 ]
Zhou, Xuda [1 ]
Zhou, Xuehai [1 ]
Wang, Chao [1 ]
Li, Xi [1 ]
机构
[1] USTC, Dept Comp Sci & Technol, Hefei, Peoples R China
基金
美国国家科学基金会;
关键词
accelerator; deep neural networks; FPGA; pipeline; inference;
D O I
10.1109/ICPADS.2016.139
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
It has been a new research hot topic to speed up the inference process of deep neural networks (DNNs) by hardware accelerators based on field programmable gate arrays (FPGAs). Because of the layer-wise structure and data dependency between layers, previous studies commonly focus on the inherent parallelism of a single layer to reduce the computation time but neglect the parallelism between layers. In this paper, we propose a pipeline energy-efficient accelerator named PIE to accelerate the DNN inference computation by pipelining two adjacent layers. Through realizing two adjacent layers in different calculation orders, the data dependency between layers can be weakened. As soon as a layer produces an output, the next layer reads the output as an input and starts the parallel computation immediately in another calculation method. In such a way, computations between adjacent layers are pipelined. We conduct our experiments on a Zedboard development kit using Xilinx Zynq-7000 FPGA, compared with Intel Core i7 4.0GHz CPU and NVIDIA K40C GPU. Experimental results indicate that PIE is 4.82x faster than CPU and can reduce the energy consumptions of CPU and GPU by 355.35x and 12.02x respectively. Besides, compared with the none-pipelined method that layers are processed in serial, PIE improves the performance by nearly 50%.
引用
收藏
页码:1067 / 1074
页数:8
相关论文
共 50 条
  • [21] An Efficient Accelerator for Deep Convolutional Neural Networks
    Kuo, Yi-Xian
    Lai, Yeong-Kang
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS - TAIWAN (ICCE-TAIWAN), 2020,
  • [22] Enabling Energy-Efficient Inference for Self-Attention Mechanisms in Neural Networks
    Chen, Qinyu
    Sun, Congyi
    Lu, Zhonghai
    Gao, Chang
    [J]. 2022 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2022): INTELLIGENT TECHNOLOGY IN THE POST-PANDEMIC ERA, 2022, : 25 - 28
  • [23] Inference and Energy Efficient Design of Deep Neural Networks for Embedded Devices
    Galanis, Ioannis
    Anagnostopoulos, Iraklis
    Nguyen, Chinh
    Bares, Guillermo
    Burkard, Dona
    [J]. 2020 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI 2020), 2020, : 36 - 41
  • [24] TRAINING DEEP SPIKING NEURAL NETWORKS FOR ENERGY-EFFICIENT NEUROMORPHIC COMPUTING
    Srinivasan, Gopalakrishnan
    Lee, Chankyu
    Sengupta, Abhronil
    Panda, Priyadarshini
    Sarwar, Syed Shakib
    Roy, Kaushik
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 8549 - 8553
  • [25] Spiking Deep Convolutional Neural Networks for Energy-Efficient Object Recognition
    Cao, Yongqiang
    Chen, Yang
    Khosla, Deepak
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2015, 113 (01) : 54 - 66
  • [26] Pruning Deep Neural Networks for Green Energy-Efficient Models: A Survey
    Tmamna, Jihene
    Ben Ayed, Emna
    Fourati, Rahma
    Gogate, Mandar
    Arslan, Tughrul
    Hussain, Amir
    Ayed, Mounir Ben
    [J]. COGNITIVE COMPUTATION, 2024,
  • [27] Spiking Deep Convolutional Neural Networks for Energy-Efficient Object Recognition
    Yongqiang Cao
    Yang Chen
    Deepak Khosla
    [J]. International Journal of Computer Vision, 2015, 113 : 54 - 66
  • [28] A Pipelined Energy-efficient Hardware Accelaration for Deep Convolutional Neural Networks
    Alaeddine, Hmidi
    Jihene, Malek
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON DESIGN & TEST OF INTEGRATED MICRO & NANO-SYSTEMS (DTS), 2019,
  • [29] Cambricon-G: A Polyvalent Energy-Efficient Accelerator for Dynamic Graph Neural Networks
    Song, Xinkai
    Zhi, Tian
    Fan, Zhe
    Zhang, Zhenxing
    Zeng, Xi
    Li, Wei
    Hu, Xing
    Du, Zidong
    Guo, Qi
    Chen, Yunji
    [J]. IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (01) : 116 - 128
  • [30] EnGN: A High-Throughput and Energy-Efficient Accelerator for Large Graph Neural Networks
    Liang, Shengwen
    Wang, Ying
    Liu, Cheng
    He, Lei
    Li, Huawei
    Xu, Dawen
    Li, Xiaowei
    [J]. IEEE TRANSACTIONS ON COMPUTERS, 2021, 70 (09) : 1511 - 1525