Impact of Phase-Change Memory Drift on Energy Efficiency and Accuracy of Analog Compute-in-Memory Deep Learning Inference (Invited)

被引:4
|
作者
Frank, Martin M. [1 ]
Li, Ning [1 ]
Rasch, Malte J. [1 ]
Jain, Shubham [1 ]
Chen, Ching-Tzu [1 ]
Muralidhar, Ramachandran [1 ]
Han, Jin-Ping [1 ]
Narayanan, Vijay [1 ]
Philip, Timothy M. [2 ]
Brew, Kevin [2 ]
Simon, Andrew [2 ]
Saraf, Iqbal [2 ]
Saulnier, Nicole [2 ]
Boybat, Irem [3 ]
Wozniak, Stanislaw [3 ]
Sebastian, Abu [3 ]
Narayanan, Pritish [4 ]
Mackin, Charles [4 ]
Chen, An [4 ]
Tsai, Hsinyu [4 ]
Burr, Geoffrey W. [4 ]
机构
[1] IBM TJ Watson Res Ctr, Yorktown Hts, NY 10598 USA
[2] IBM Res Albany, Albany, NY USA
[3] IBM Res Zurich, Ruschlikon, Switzerland
[4] IBM Res Almaden, San Jose, CA USA
关键词
Analog memory; in-memory computing; neural networks; phase change memory; semiconductor device reliability; ARRAY;
D O I
10.1109/IRPS48203.2023.10117874
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
Among the emerging approaches for deep learning acceleration, compute-in-memory (CIM) in crossbar arrays, in conjunction with optimized digital computation and communication, is attractive for achieving high execution speeds and energy efficiency. Analog phase-change memory (PCM) is particularly promising for this purpose. However, resistance typically drifts, which can degrade deep learning accuracy over time. Herein, we first discuss drift and noise mitigation by integrating projection liners into analog mushroom-type PCM devices, as well as tradeoffs with dynamic range. We then study their impact on inference accuracy for the Transformer-based language model BERT. We find that accuracy loss after extended drift can be minimal with an optimized mapping of weights to cells comprising two pairs of liner PCM devices of varying significance. Finally, we address the impact of drift on energy consumption during inference through a combination of drift, circuit, and architecture simulations. For a range of typical drift coefficients, we show that the peak vector-matrix multiplication (VMM) energy efficiency of a recently proposed heterogeneous CIM accelerator in 14 nm technology can increase by 3% to 15% over the course of one day to ten years. For convolutional neural network (CNN), long short-term memory (LSTM) and Transformer benchmarks, the increase in sustained energy efficiency remains below 10%, being greatest for models dominated by analog computation. Longer VMM integration times increase the energy impact of drift.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Accuracy and Resiliency of Analog Compute-in-Memory Inference Engines
    Wan, Zhe
    Wang, Tianyi
    Zhou, Yiming
    Iyer, Subramanian S.
    Roychowdhury, Vwani P.
    ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2022, 18 (02)
  • [2] Analog Compute-in-Memory For AI Edge Inference
    Fick, D.
    2022 INTERNATIONAL ELECTRON DEVICES MEETING, IEDM, 2022,
  • [3] Impact of Phase-Change Memory Flicker Noise and Weight Drift on Analog Hardware Inference for Large-Scale Deep Learning Networks
    Han, Jin-Ping
    Rasch, Malte J.
    Liu, Zuoguang
    Solomon, Paul
    Brew, Kevin
    Cheng, Kangguo
    Ok, Injo
    Chan, Victor
    Longstreet, Michael
    Kim, Wanki
    Bruce, Robert L.
    Cheng, Cheng-Wei
    Saulnier, Nicole
    Narayanan, Vijay
    ADVANCED INTELLIGENT SYSTEMS, 2022, 4 (05)
  • [4] PCM-Based Analog Compute-In-Memory: Impact of Device Non-Idealities on Inference Accuracy
    Sun, X.
    Khwa, W. S.
    Chen, Y. S.
    Lee, C. H.
    Lee, H. Y.
    Yu, S. M.
    Naous, R.
    Wu, J. Y.
    Chen, T. C.
    Bao, X.
    Chang, M. F.
    Diaz, C. H.
    Wong, H-S P.
    Akarvardar, K.
    IEEE TRANSACTIONS ON ELECTRON DEVICES, 2021, 68 (11) : 5585 - 5591
  • [5] Phase-Change Memory Models for Deep Learning Training and Inference
    Nandakumar, S. R.
    Boybat, Irem
    Joshi, Vinay
    Piveteau, Christophe
    Le Gallo, Manuel
    Rajendran, Bipin
    Sebastian, Abu
    Eleftheriou, Evangelos
    2019 26TH IEEE INTERNATIONAL CONFERENCE ON ELECTRONICS, CIRCUITS AND SYSTEMS (ICECS), 2019, : 727 - 730
  • [6] Compute-In-Memory Technologies for Deep Learning Acceleration
    Meng, Fan-husan
    Lu, Wei D.
    IEEE NANOTECHNOLOGY MAGAZINE, 2024, 18 (01) : 44 - 52
  • [7] Compute-in-Memory Technologies and Architectures for Deep Learning Workloads
    Ali, Mustafa
    Roy, Sourjya
    Saxena, Utkarsh
    Sharma, Tanvi
    Raghunathan, Anand
    Roy, Kaushik
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2022, 30 (11) : 1615 - 1630
  • [8] CoMeFa: Deploying Compute-in-Memory on FPGAs for Deep Learning Acceleration
    Arora, Aman
    Bhamburkar, Atharva
    Borda, Aatman
    Anand, Tanmay
    Sehgal, Rishabh
    Hanindhito, Bagus
    Gaillardon, Pierre-Emmanuel
    Kulkarni, Jaydeep
    John, Lizy K.
    ACM TRANSACTIONS ON RECONFIGURABLE TECHNOLOGY AND SYSTEMS, 2023, 16 (03)
  • [9] Precision of synaptic weights programmed in phase-change memory devices for deep learning inference
    Nandakumar, S. R.
    Boybat, Irem
    Han, Jin-Ping
    Ambrogio, Stefano
    Adusumilli, Praneet
    Bruce, Robert L.
    BrightSky, Matthew
    Rasch, Malte
    Le Gallo, Manuel
    Sebastian, Abu
    2020 IEEE INTERNATIONAL ELECTRON DEVICES MEETING (IEDM), 2020,
  • [10] Compute-in-Memory Chips for Deep Learning: Recent Trends and Prospects
    Yu, Shimeng
    Jiang, Hongwu
    Huang, Shanshi
    Peng, Xiaochen
    Lu, Anni
    IEEE CIRCUITS AND SYSTEMS MAGAZINE, 2021, 21 (03) : 31 - 56