Optimization of DRAM based PIM Architecture for Energy-Efficient Deep Neural Network Training

被引:3
|
作者
Sudarshan, Chirag [1 ]
Sadi, Mohammad Hassani [1 ]
Weis, Christian [1 ]
Wehn, Norbert [1 ]
机构
[1] Tech Univ Kaiserslautern, Microelect Syst Design Res Grp, Kaiserslautern, Germany
关键词
Processing-in-Memory; PIM; Compute-in-Memory; DRAM; DNN training; MEMORY;
D O I
10.1109/ISCAS48785.2022.9937832
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Deep Neural Network (DNN) training consumes high-energy. On the other hand, DNNs deployed on edge devices demand very high-energy efficiency. In this context, Processing-in-Memory (PIM) is an emerging compute paradigm that bridges the memory-computation gap to improve the energy-efficiency. DRAMs are one such memory type employed for designing energy-efficient PIM architectures for DNN training. One of the major issues of DRAM-PIM architectures designed for DNN training is the high number of internal data accesses within a bank between the memory arrays and the PIM computation units (e.g. 51% more than inference). These internal data accesses in the state-of-the-art DRAM PIM architectures consume very high energy compared to computation units. Hence, it is important to reduce the internal data access energy within the DRAM bank for further improving the energy efficiency of DRAM-PIM architectures. We present three novel optimizations that together reduce the internal data access energy up to 81.54%. Our first optimization modifies the bank data access circuit to enable partial accesses of data instead of the conventional fixed granularity accesses, thereby exploiting the available sparsity during training. The second optimization is to have a dedicated low-energy region within the DRAM bank that has low capacitive load of global wires and shorter data movement. Finally, we propose a 12-bit high dynamic range floating-point format called TinyFloat that reduces the total number of data access energy by 20% compared to IEEE 754 half and single precision.
引用
收藏
页码:1472 / 1476
页数:5
相关论文
共 50 条
  • [41] A Proposal for Energy-Efficient Cellular Neural Network Based on Spintronic Devices
    Pan, Chenyun
    Naeemi, Azad
    [J]. IEEE TRANSACTIONS ON NANOTECHNOLOGY, 2016, 15 (05) : 820 - 827
  • [42] An Energy-Efficient Accelerator for Rain Removal Based on Convolutional Neural Network
    Rao, Lei
    Zhang, Bin
    Zhao, Jizhong
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2021, 68 (08) : 2957 - 2961
  • [43] Rank order coding based spiking convolutional neural network architecture with energy-efficient membrane voltage updates
    Tang, Hoyoung
    Cho, Donghyeon
    Lew, Dongwoo
    Kim, Taehwan
    Park, Jongsun
    [J]. NEUROCOMPUTING, 2020, 407 : 300 - 312
  • [44] Energy-Efficient Network Architecture for Smart City Development
    Jayadeyan, Ashwini
    Syafiza, Ida
    Ahmad, Mohd Riduan
    Azyze, Nur Latif Azyze Mohd Shaari
    Rasidi, Norasida Farahizzaty
    Zainan, Nur Hidayah
    [J]. PRZEGLAD ELEKTROTECHNICZNY, 2023, 99 (09): : 135 - 140
  • [45] Memory Efficient Deep Neural Network Training
    Shilova, Alena
    [J]. EURO-PAR 2021: PARALLEL PROCESSING WORKSHOPS, 2022, 13098 : 515 - 519
  • [46] Optimization of the Energy-Efficient Relay-Based Massive IoT Network
    Lv, Tiejun
    Lin, Zhipeng
    Huang, Pingmu
    Zeng, Jie
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2018, 5 (04): : 3043 - 3058
  • [47] C-DNN: An Energy-Efficient Complementary Deep-Neural-Network Processor With Heterogeneous CNN/SNN Core Architecture
    Kim, Sangyeob
    Kim, Soyeon
    Hong, Seongyon
    Kim, Sangjin
    Han, Donghyeon
    Choi, Jiwon
    Yoo, Hoi-Jun
    [J]. IEEE JOURNAL OF SOLID-STATE CIRCUITS, 2024, 59 (01) : 157 - 172
  • [48] Energy-efficient cooperative inference via adaptive deep neural network splitting at the edge
    Labriji, Ibtissam
    Merluzzi, Mattia
    Airod, Fatima Ezzahra
    Strinati, Emilio Calvanese
    [J]. ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 1712 - 1717
  • [49] IMC: Energy-Efficient In-Memory Convolver for Accelerating Binarized Deep Neural Network
    Angizi, Shaahin
    Fan, Deliang
    [J]. PROCEEDINGS OF NEUROMORPHIC COMPUTING SYMPOSIUM (NCS 2017), 2017,
  • [50] Energy-Efficient and High-Performance NoC Architecture and Mapping Solution for Deep Neural Networks
    Reza, Md Farhadur
    Ampadu, Paul
    [J]. PROCEEDINGS OF THE 13TH IEEE/ACM INTERNATIONAL SYMPOSIUM ON NETWORKS-ON-CHIP (NOCS'19), 2019,