Fast and robust analog in-memory deep neural network training

被引:0
|
作者
Rasch, Malte J. [1 ,2 ]
Carta, Fabio [1 ]
Fagbohungbe, Omobayode [1 ]
Gokmen, Tayfun [1 ]
机构
[1] IBM Res, TJ Watson Res Ctr, Yorktown Hts, NY 10598 USA
[2] Sony AI, Zurich, Switzerland
关键词
DEVICES; CHIP;
D O I
10.1038/s41467-024-51221-z
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Analog in-memory computing is a promising future technology for efficiently accelerating deep learning networks. While using in-memory computing to accelerate the inference phase has been studied extensively, accelerating the training phase has received less attention, despite its arguably much larger compute demand to accelerate. While some analog in-memory training algorithms have been suggested, they either invoke significant amount of auxiliary digital compute-accumulating the gradient in digital floating point precision, limiting the potential speed-up-or suffer from the need for near perfectly programming reference conductance values to establish an algorithmic zero point. Here, we propose two improved algorithms for in-memory training, that retain the same fast runtime complexity while resolving the requirement of a precise zero point. We further investigate the limits of the algorithms in terms of conductance noise, symmetry, retention, and endurance which narrow down possible device material choices adequate for fast and robust in-memory deep neural network training. Analog in-memory computing recent hardware implementations focused mainly on accelerating inference deployment. In this work, to improve the training process, the authors propose algorithms for supervised training of deep neural networks on analog in-memory AI accelerator hardware.
引用
收藏
页数:15
相关论文
共 50 条
  • [41] A scalable and reconfigurable in-memory architecture for ternary deep spiking neural network with ReRAM based neurons
    Lin, Jie
    Yuan, Jiann-Shiun
    NEUROCOMPUTING, 2020, 375 : 102 - 112
  • [42] EPMC: efficient parallel memory compression in deep neural network training
    Zailong Chen
    Shenghong Yang
    Chubo Liu
    Yikun Hu
    Kenli Li
    Keqin Li
    Neural Computing and Applications, 2022, 34 : 757 - 769
  • [43] EPMC: efficient parallel memory compression in deep neural network training
    Chen, Zailong
    Yang, Shenghong
    Liu, Chubo
    Hu, Yikun
    Li, Kenli
    Li, Keqin
    NEURAL COMPUTING & APPLICATIONS, 2022, 34 (01): : 757 - 769
  • [44] An In-Memory Computing SRAM Macro for Memory-Augmented Neural Network
    Kim, Sunghoon
    Lee, Wonjae
    Kim, Sundo
    Park, Sungjin
    Jeon, Dongsuk
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2022, 69 (03) : 1687 - 1691
  • [45] Multilevel artificial electronic synaptic device of direct grown robust MoS2 based memristor array for in-memory deep neural network
    Naqi, Muhammad
    Kang, Min Seok
    Liu, Na
    Kim, Taehwan
    Baek, Seungho
    Bala, Arindam
    Moon, Changgyun
    Park, Jongsun
    Kim, Sunkook
    NPJ 2D MATERIALS AND APPLICATIONS, 2022, 6 (01)
  • [46] Multilevel artificial electronic synaptic device of direct grown robust MoS2 based memristor array for in-memory deep neural network
    Muhammad Naqi
    Min Seok Kang
    Na liu
    Taehwan Kim
    Seungho Baek
    Arindam Bala
    Changgyun Moon
    Jongsun Park
    Sunkook Kim
    npj 2D Materials and Applications, 6
  • [47] Current Status and Issues of in-memory Accelerators for Deep Neural Networks
    Deguchi, Jun
    2021 INTERNATIONAL SYMPOSIUM ON VLSI DESIGN, AUTOMATION AND TEST (VLSI-DAT), 2021,
  • [48] Vesti: An In-Memory Computing Processor for Deep Neural Networks Acceleration
    Jiang, Zhewei
    Yin, Shihui
    Kim, Minkyu
    Gupta, Tushar
    Seok, Mingoo
    Seo, Jae-sun
    CONFERENCE RECORD OF THE 2019 FIFTY-THIRD ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, 2019, : 1516 - 1521
  • [49] Enabling Secure NVM-Based in-Memory Neural Network Computing by Sparse Fast Gradient Encryption
    Cai, Yi
    Chen, Xiaoming
    Tian, Lu
    Wang, Yu
    Yang, Huazhong
    IEEE TRANSACTIONS ON COMPUTERS, 2020, 69 (11) : 1596 - 1610
  • [50] Robust algorithm for neural network training
    Manic, M
    Wilamowski, B
    PROCEEDING OF THE 2002 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-3, 2002, : 1528 - 1533