Fast and robust analog in-memory deep neural network training

被引:0
|
作者
Rasch, Malte J. [1 ,2 ]
Carta, Fabio [1 ]
Fagbohungbe, Omobayode [1 ]
Gokmen, Tayfun [1 ]
机构
[1] IBM Res, TJ Watson Res Ctr, Yorktown Hts, NY 10598 USA
[2] Sony AI, Zurich, Switzerland
关键词
DEVICES; CHIP;
D O I
10.1038/s41467-024-51221-z
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Analog in-memory computing is a promising future technology for efficiently accelerating deep learning networks. While using in-memory computing to accelerate the inference phase has been studied extensively, accelerating the training phase has received less attention, despite its arguably much larger compute demand to accelerate. While some analog in-memory training algorithms have been suggested, they either invoke significant amount of auxiliary digital compute-accumulating the gradient in digital floating point precision, limiting the potential speed-up-or suffer from the need for near perfectly programming reference conductance values to establish an algorithmic zero point. Here, we propose two improved algorithms for in-memory training, that retain the same fast runtime complexity while resolving the requirement of a precise zero point. We further investigate the limits of the algorithms in terms of conductance noise, symmetry, retention, and endurance which narrow down possible device material choices adequate for fast and robust in-memory deep neural network training. Analog in-memory computing recent hardware implementations focused mainly on accelerating inference deployment. In this work, to improve the training process, the authors propose algorithms for supervised training of deep neural networks on analog in-memory AI accelerator hardware.
引用
收藏
页数:15
相关论文
共 50 条
  • [31] Soft Memory Box: A Virtual Shared Memory Framework for Fast Deep Neural Network Training in Distributed High Performance Computing
    Ahn, Shinyoung
    Kim, Joongheon
    Lim, Eunji
    Kang, Sungwon
    IEEE ACCESS, 2018, 6 : 26493 - 26504
  • [32] Robust feature space separation for deep convolutional neural network training
    Sekmen A.
    Parlaktuna M.
    Abdul-Malek A.
    Erdemir E.
    Koku A.B.
    Discover Artificial Intelligence, 2021, 1 (01):
  • [33] An MRAM-based Deep In-Memory Architecture for Deep Neural Networks
    Patil, Ameya D.
    Hua, Haocheng
    Gonugondla, Sujan
    Kang, Mingu
    Shanbhag, Naresh R.
    2019 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2019,
  • [34] GEMINI: Fast Failure Recovery in Distributed Training with In-Memory Checkpoints
    Wang, Zhuang
    Jia, Zhen
    Zheng, Shuai
    Zhang, Zhen
    Fu, Xinwei
    Ng, T. S. Eugene
    Wang, Yida
    PROCEEDINGS OF THE TWENTY-NINTH ACM SYMPOSIUM ON OPERATING SYSTEMS PRINCIPLES, SOSP 2023, 2023, : 364 - 381
  • [35] Hadamard product-based in-memory computing design for floating point neural network training
    Fan, Anjunyi
    Fu, Yihan
    Tao, Yaoyu
    Jin, Zhonghua
    Han, Haiyue
    Liu, Huiyu
    Zhang, Yaojun
    Yan, Bonan
    Yang, Yuchao
    Huang, Ru
    NEUROMORPHIC COMPUTING AND ENGINEERING, 2023, 3 (01):
  • [36] ALPINE: Analog In-Memory Acceleration With Tight Processor Integration for Deep Learning
    Klein, Joshua
    Boybat, Irem
    Qureshi, Yasir Mahmood
    Dazzi, Martino
    Levisse, Alexandre
    Ansaloni, Giovanni
    Zapater, Marina
    Sebastian, Abu
    Atienza, David
    IEEE TRANSACTIONS ON COMPUTERS, 2023, 72 (07) : 1985 - 1998
  • [37] NNPIM: A Processing In-Memory Architecture for Neural Network Acceleration
    Gupta, Saransh
    Imani, Mohsen
    Kaur, Harveen
    Rosing, Tajana Simunic
    IEEE TRANSACTIONS ON COMPUTERS, 2019, 68 (09) : 1325 - 1337
  • [38] Fast training algorithm for deep neural network using multiple GPUs
    Dai, L. (lrdai@ustc.edu.cn), 1600, Tsinghua University (53):
  • [39] Fast Deep Neural Network Training on Distributed Systems and Cloud TPUs
    You, Yang
    Zhang, Zhao
    Hsieh, Cho-Jui
    Demmel, James
    Keutzer, Kurt
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2019, 30 (11) : 2449 - 2462
  • [40] SIMBA: A Skyrmionic In-Memory Binary Neural Network Accelerator
    Miriyala, Venkata Pavan Kumar
    Vishwanath, Kale Rahul
    Fong, Xuanyao
    IEEE TRANSACTIONS ON MAGNETICS, 2020, 56 (11)