A Quantized Training Framework for Robust and Accurate ReRAM-based Neural Network Accelerators

被引:7
|
作者
Zhang, Chenguang [1 ]
Zhou, Pingqiang [1 ]
机构
[1] ShanghaiTech Univ, Sch Informat Sci & Technol, Shanghai, Peoples R China
关键词
ReRAM; Neural Network; Variation; Robust; Quantize;
D O I
10.1145/3394885.3431528
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Neural networks (NN), especially deep neural networks (DNN), have achieved great success in lots of fields. ReRAM crossbar, as a promising candidate, is widely employed to accelerate neural network owing to its nature of processing MVM. However, ReRAM crossbar suffers high conductance variation due to many non-ideal effects, resulting in great inference accuracy degradation. Recent works use uniform quantization to enhance the tolerance of conductance variation, but these methods still suffer high accuracy loss with large variation. In this paper, firstly, we analyze the impact of the quantization and conductance variation on the accuracy. Then, based on two observation, we propose a quantized training framework to enhance the robustness and accuracy of the neural network running on the accelerator, by introducing a smart non-uniform quantizer. This framework consists of a robust trainable quantizer and a corresponding training method, and needs no extra hardware overhead and compatible with a standard neural network training procedure. Experimental results show that our proposed method can improve inference accuracy by 10% similar to 30% under large variation, compared with uniform quantization method.
引用
收藏
页码:43 / 48
页数:6
相关论文
共 50 条
  • [1] Offline Training-Based Mitigation of IR Drop for ReRAM-Based Deep Neural Network Accelerators
    Lee, Sugil
    Fouda, Mohammed E.
    Lee, Jongeun
    Eltawil, Ahmed M.
    Kurdahi, Fadi
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (02) : 521 - 532
  • [2] Trained Biased Number Representation for ReRAM-Based Neural Network Accelerators
    Wang, Weijia
    Lin, Bill
    ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2019, 15 (02)
  • [3] A Unified Framework for Training, Mapping and Simulation of ReRAM-Based Convolutional Neural Network Acceleration
    Liu, He
    Han, Jianhui
    Zhang, Youhui
    IEEE COMPUTER ARCHITECTURE LETTERS, 2019, 18 (01) : 63 - 66
  • [4] Quarry: Quantization-based ADC Reduction for ReRAM-based Deep Neural Network Accelerators
    Azamat, Azat
    Asim, Faaiz
    Lee, Jongeun
    2021 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER AIDED DESIGN (ICCAD), 2021,
  • [5] Partial Sum Quantization for Reducing ADC Size in ReRAM-Based Neural Network Accelerators
    Azamat, Azat
    Asim, Faaiz
    Kim, Jintae
    Lee, Jongeun
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (12) : 4897 - 4908
  • [6] Mathematical Framework for Optimizing Crossbar Allocation for ReRAM-based CNN Accelerators
    Li, Wanqian
    Han, Yinhe
    Chen, Xiaoming
    ACM TRANSACTIONS ON DESIGN AUTOMATION OF ELECTRONIC SYSTEMS, 2024, 29 (01)
  • [7] Design Framework for ReRAM-Based DNN Accelerators with Accuracy and Hardware Evaluation
    Kao, Hsu-Yu
    Huang, Shih-Hsu
    Cheng, Wei-Kai
    ELECTRONICS, 2022, 11 (13)
  • [8] A Quantized Training Method to Enhance Accuracy of ReRAM-based Neuromorphic Systems
    Yang, Qing
    Li, Hai
    Wu, Qing
    2018 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2018,
  • [9] Hardware attacks on ReRAM-based AI accelerators
    Heidary, Masoud
    Joardar, Biresh Kumar
    17TH IEEE DALLAS CIRCUITS AND SYSTEMS CONFERENCE, DCAS 2024, 2024,
  • [10] FARe: Fault-Aware GNN Training on ReRAM-based PIM Accelerators
    Dhingra, Pratyush
    Ogbogu, Chukwufumnanya
    Joardar, Biresh Kumar
    Doppa, Janardhan Rao
    Kalyanaraman, Ananth
    Pande, Partha Pratim
    2024 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION, DATE, 2024,