An Energy-Efficient Deep Learning Processor with Heterogeneous Multi-Core Architecture for Convolutional Neural Networks and Recurrent Neural Networks

被引:0
|
作者
Shin, Dongjoo [1 ]
Lee, Jinmook [1 ]
Lee, Jinsu [1 ]
Lee, Juhyoung [1 ]
Yoo, Hoi-Jun [1 ]
机构
[1] Korea Adv Inst Sci & Technol, Sch Elect Engn, 291 Daehak Ro, Daejeon, South Korea
关键词
deep learning; convolutional neural network; recurrent neural network; heterogeneous; LUT;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
An energy-efficient deep learning processor is proposed for convolutional neural networks (CNNs) and recurrent neural networks (RNNs) in mobile platforms. The 16mm(2) chip is fabricated using 65nm technology with 3 key features, 1) Reconfigurable heterogeneous architecture to support both CNNs and RNNs, 2) LUT-based reconfigurable multiplier optimized for dynamic fixed-point with the on-line adaptation, 3) Quantization table-based matrix multiplication to reduce off-chip memory access and remove duplicated multiplications. As a result, compared to the [2] and [3], this work shows 20x and 4.5x higher energy efficiency, respectively. Also, DNPU shows 6.5(x) higher energy efficiency compared to the [5].
引用
收藏
页数:2
相关论文
共 50 条
  • [1] DNPU: An Energy-Efficient Deep-Learning Processor with Heterogeneous Multi-Core Architecture
    Shin, Dongjoo
    Lee, Jinmook
    Lee, Jinsu
    Lee, Juhyoung
    Yoo, Hoi-Jun
    [J]. IEEE MICRO, 2018, 38 (05) : 85 - 93
  • [2] A Heterogeneous and Reconfigurable Embedded Architecture for Energy-Efficient Execution of Convolutional Neural Networks
    Luebeck, Konstantin
    Bringmann, Oliver
    [J]. ARCHITECTURE OF COMPUTING SYSTEMS - ARCS 2019, 2019, 11479 : 267 - 280
  • [3] SLIT: An Energy-Efficient Reconfigurable Hardware Architecture for Deep Convolutional Neural Networks
    Tran, Thi Diem
    Nakashima, Yasuhiko
    [J]. IEICE TRANSACTIONS ON ELECTRONICS, 2021, E104C (07) : 319 - 329
  • [4] An Energy-Efficient Architecture for Binary Weight Convolutional Neural Networks
    Wang, Yizhi
    Lin, Jun
    Wang, Zhongfeng
    [J]. IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2018, 26 (02) : 280 - 293
  • [5] An Energy-Efficient Deep Belief Network Processor Based on Heterogeneous Multi-Core Architecture With Transposable Memory and On-Chip Learning
    Wu, Jiajun
    Huang, Xuan
    Yang, Le
    Wang, Jipeng
    Liu, Bingqiang
    Wen, Ziyuan
    Li, Juhui
    Yu, Guoyi
    Chong, Kwen-Siong
    Wang, Chao
    [J]. IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2021, 11 (04) : 725 - 738
  • [6] Energy-Efficient Convolutional Neural Networks via Recurrent Data Reuse
    Mocerino, Luca
    Tenace, Valerio
    Calimera, Andrea
    [J]. 2019 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE), 2019, : 848 - 853
  • [7] An energy-efficient deep convolutional neural networks coprocessor for multi-object detection
    Wu, Yuancong
    Wang, J. J.
    Qian, Kun
    Liu, Yanchen
    Guo, Rui
    Hu, S. G.
    Yu, Q.
    Chen, T. P.
    Liu, Y.
    Rong, Limei
    [J]. MICROELECTRONICS JOURNAL, 2020, 98
  • [8] Eyeriss: A Spatial Architecture for Energy-Efficient Dataflow for Convolutional Neural Networks
    Chen, Yu-Hsin
    Emer, Joel
    Sze, Vivienne
    [J]. 2016 ACM/IEEE 43RD ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA), 2016, : 367 - 379
  • [9] Energy-Efficient Architecture for FPGA-based Deep Convolutional Neural Networks with Binary Weights
    Duan, Yunzhi
    Li, Shuai
    Zhang, Ruipeng
    Wang, Qi
    Chen, Jienan
    Sobelman, Gerald E.
    [J]. 2018 IEEE 23RD INTERNATIONAL CONFERENCE ON DIGITAL SIGNAL PROCESSING (DSP), 2018,
  • [10] COSY: An Energy-Efficient Hardware Architecture for Deep Convolutional Neural Networks based on Systolic Array
    Yin, Chen
    Chen, Qiang
    Tian, Miren
    Ji, Mohan
    Zou, Chenglong
    Wang, Yin'an
    Wang, Bo
    [J]. 2017 IEEE 23RD INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS (ICPADS), 2017, : 180 - 189