HD-CIM: Hybrid-Device Computing-In-Memory Structure Based on MRAM and SRAM to Reduce Weight Loading Energy of Neural Networks

被引:12
|
作者
Zhang, He [1 ]
Liu, Junzhan [1 ]
Bai, Jinyu [1 ]
Li, Sai [2 ]
Luo, Lichuan [1 ]
Wei, Shaoqian [1 ]
Wu, Jianxin [1 ]
Kang, Wang [1 ]
机构
[1] Beihang Univ, Sch Integrated Circuit Sci & Engn, Fert Beijing Inst, Beijing 100191, Peoples R China
[2] Beihang Univ, Shen Yuan Honors Coll, Sch Integrated Circuit Sci & Engn, Fert Beijing Inst, Beijing 100191, Peoples R China
关键词
Computing-in-memory (CIM); neural networks (NNs); MRAM; SRAM; EFFICIENT; DESIGN; MACRO; ACCELERATORS; HARDWARE; 6T-SRAM;
D O I
10.1109/TCSI.2022.3199440
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
SRAM based computing-in-memory (SRAM-CIM) techniques have been widely studied for neural networks (NNs) to solve the "Von Neumann bottleneck". However, as the scale of the NN model increasingly expands, the weight cannot be fully stored on-chip owing to the big device size (limited capacity) of SRAM. In this case, the NN weight data have to be frequently loaded from external memories, such as DRAM and Flash memory, which results in high energy consumption and low efficiency. In this paper, we propose a hybrid-device computing-in-memory (HD-CIM) architecture based on SRAM and MRAM (magnetic random-access memory). In our HD-CIM, the NN weight data are stored in on-chip MRAM and are loaded into SRAM-CIM core, significantly reducing energy and latency. Besides, in order to improve the data transfer efficiency between MRAM and SRAM, a high-speed pipelined MRAM readout structure is proposed to reduce the BL charging time. Our results show that the NN weight data loading energy in our design is only 0.242 pJ/bit, which is 289 x less in comparison with that from off-chip DRAM. Moreover, the energy breakdown and efficiency are analyzed based on different NN models, such as VGG19, ResNetl8 and MobileNetVl. Our design can improve 58 x to 124 x energy efficiency.
引用
收藏
页码:4465 / 4474
页数:10
相关论文
共 4 条
  • [1] A negative capacitance FET based energy efficient 6T SRAM computing-in-memory (CiM) cell design for deep neural networks
    Birudu, Venu
    Yellampalli, Siva Sankar
    Vaddi, Ramesh
    MICROELECTRONICS JOURNAL, 2023, 139
  • [2] Scale-CIM: Precision-scalable computing-in-memory for energy-efficient quantized neural networks
    Lee, Young Seo
    Gong, Young -Ho
    Chung, Sung Woo
    JOURNAL OF SYSTEMS ARCHITECTURE, 2023, 134
  • [3] MixCIM: A Hybrid-Cell-Based Computing-in-Memory Macro with Less-Data-Movement and Activation-Memory-Reuse for Depthwise Separable Neural Networks
    Qiao, Xin
    Song, Jiahao
    Yang, Youming
    Wei, Renjie
    Tang, Xiyuan
    Li, Meng
    Wang, Runsheng
    Wang, Yuan
    2024 IEEE CUSTOM INTEGRATED CIRCUITS CONFERENCE, CICC, 2024,
  • [4] A Novel Small-Signal Ferroelectric Memcapacitor based Capacitive Computing-In-Memory for Area- and Energy-Efficient Quantized Neural Networks
    Xu, Weikai
    Luo, Jin
    Fu, Boyi
    Fu, Zhiyuan
    Wang, Kaifeng
    Su, Chang
    Huang, Qianqian
    Huang, Ru
    8TH IEEE ELECTRON DEVICES TECHNOLOGY & MANUFACTURING CONFERENCE, EDTM 2024, 2024, : 238 - 240