A DNN Compression Framework for SOT-MRAM-based Processing-In-Memory Engine

被引:4
|
作者
Yuan, Geng [1 ]
Ma, Xiaolong [1 ]
Lin, Sheng [1 ]
Li, Zhengang [1 ]
Deng, Jieren [2 ]
Ding, Caiwen [2 ]
机构
[1] Northeastern Univ, Boston, MA 02115 USA
[2] Univ Connecticut, Storrs, CT USA
关键词
NEURAL-NETWORKS; ENERGY;
D O I
10.1109/SOCC49529.2020.9524757
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
The computing wall and data movement challenges of deep neural networks (DNNs) have exposed the limitations of conventional CMOS-based DNN accelerators. Furthermore, the deep structure and large model size will make DNNs prohibitive to embedded systems and IoT devices, where low power consumption is required. To address these challenges, spin-orbit torque magnetic random-access memory (SOT-MRAM) and SOT-MRAM based Processing-In-Memory (PIM) engines have been used to reduce the power consumption of DNNs since SOT-MRAM has the characteristic of near-zero standby power, high density, non-volatile. However, the drawbacks of SOT-MRAM based PIM engines such as high writing latency and requiring low bit-width data decrease its popularity as a favorable energy-efficient DNN accelerator. To mitigate these drawbacks, we propose an ultra-energy-efficient framework by using model compression techniques including weight pruning and quantization from the software level considering the architecture of SOT-MRAM PIM. And we incorporate the alternating direction method of multipliers (ADMM) into the training phase to further guarantee the solution feasibility and satisfy SOT-MRAM hardware constraints. Thus, the footprint and power consumption of SOT-MRAM PIM can be reduced, while increasing the overall system performance rate (frame per second) in the meantime, making our proposed ADMM-based SOT-MRAM PIM more energy efficient and suitable for embedded systems or IoT devices. Our experimental results show the accuracy and compression rate of our proposed framework is consistently outperforming the reference works, while the efficiency (area & power) and performance rate of SOT-MRAM PIM engine is significantly improved.
引用
收藏
页码:37 / 42
页数:6
相关论文
共 50 条
  • [1] SOT-MRAM-Based Approximate Content Addressable Memory for DNA Classification
    Shaban, Ahmed
    Hou, Tuo-Hung
    Suri, Manan
    [J]. IEEE TRANSACTIONS ON ELECTRON DEVICES, 2024, 71 (09) : 5732 - 5738
  • [2] XBAROPT - Enabling ultra-pipelined, novel STT MRAM based processing-in-memory DNN accelerator
    Anwar, Aqeel
    Raychowdhury, Arijit
    Hatcher, Ryan
    Rakshit, Titash
    [J]. 2020 2ND IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2020), 2020, : 36 - 40
  • [3] SOT-MRAM based Analog in-Memory Computing for DNN inference
    Doevenspeck, J.
    Garello, K.
    Verhoef, B.
    Degraeve, R.
    Van Beek, S.
    Crotti, D.
    Yasin, F.
    Couet, S.
    Jayakumar, G.
    Papistas, I. A.
    Debacker, P.
    Lauwereins, R.
    Dehaene, W.
    Kar, G. S.
    Cosemans, S.
    Mallik, A.
    Verkest, D.
    [J]. 2020 IEEE SYMPOSIUM ON VLSI TECHNOLOGY, 2020,
  • [4] AlignS: A Processing-In-Memory Accelerator for DNA Short Read Alignment Leveraging SOT-MRAM
    Angizi, Shaahin
    Sun, Jiao
    Zhang, Wei
    Fan, Deliang
    [J]. PROCEEDINGS OF THE 2019 56TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2019,
  • [5] SOT-MRAM-based Binary Neural Networks Demonstration for Single Character Recognition
    Zhao, Dongyan
    Wang, Yubo
    Chen, Yanning
    Shao, Jin
    Fu, Zhen
    Liu, Fang
    Bai, Yue
    Zhao, Faqiang
    Zhong, Mingchen
    Pan, Cheng
    Dong, Yi
    Cao, Kaihua
    [J]. SPIN, 2022, 12 (01)
  • [6] A Ferroelectric FET-Based Processing-in-Memory Architecture for DNN Acceleration
    Long, Yun
    Kim, Daehyun
    Lee, Edward
    Saha, Priyabrata
    Mudassar, Burhan Ahmad
    She, Xueyuan
    Khan, Asif Islam
    Mukhopadhyay, Saibal
    [J]. IEEE JOURNAL ON EXPLORATORY SOLID-STATE COMPUTATIONAL DEVICES AND CIRCUITS, 2019, 5 (02): : 113 - 122
  • [7] Exploring a SOT-MRAM Based In-Memory Computing for Data Processing
    He, Zhezhi
    Zhang, Yang
    Angizi, Shaahin
    Gong, Boqing
    Fan, Deliang
    [J]. IEEE TRANSACTIONS ON MULTI-SCALE COMPUTING SYSTEMS, 2018, 4 (04): : 676 - 685
  • [8] SOT-MRAM-Based Design for Energy-Efficient and Reliable Binary Neural Network Acceleration
    Shaban, Ahmed
    Gothalyan, Shreshtha
    Hou, Tuo-Hung
    Suri, Manan
    [J]. IEEE TRANSACTIONS ON ELECTRON DEVICES, 2024, 71 (09) : 5367 - 5374
  • [9] SOT-MRAM-Based LUT Cell Design for Area- and Energy-Efficient FPGA
    Kim, Dongsu
    Kim, Taehwan
    Jang, Yunho
    Park, Jongsun
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2024, 71 (09) : 4276 - 4280
  • [10] A Design Framework for Processing-In-Memory Accelerator
    Gao, Di
    Shen, Tianhao
    Zhuo, Cheng
    [J]. 2018 ACM/IEEE INTERNATIONAL WORKSHOP ON SYSTEM LEVEL INTERCONNECT PREDICTION (SLIP), 2018,