A Quantized Training Framework for Robust and Accurate ReRAM-based Neural Network Accelerators

被引:7
|
作者
Zhang, Chenguang [1 ]
Zhou, Pingqiang [1 ]
机构
[1] ShanghaiTech Univ, Sch Informat Sci & Technol, Shanghai, Peoples R China
关键词
ReRAM; Neural Network; Variation; Robust; Quantize;
D O I
10.1145/3394885.3431528
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Neural networks (NN), especially deep neural networks (DNN), have achieved great success in lots of fields. ReRAM crossbar, as a promising candidate, is widely employed to accelerate neural network owing to its nature of processing MVM. However, ReRAM crossbar suffers high conductance variation due to many non-ideal effects, resulting in great inference accuracy degradation. Recent works use uniform quantization to enhance the tolerance of conductance variation, but these methods still suffer high accuracy loss with large variation. In this paper, firstly, we analyze the impact of the quantization and conductance variation on the accuracy. Then, based on two observation, we propose a quantized training framework to enhance the robustness and accuracy of the neural network running on the accelerator, by introducing a smart non-uniform quantizer. This framework consists of a robust trainable quantizer and a corresponding training method, and needs no extra hardware overhead and compatible with a standard neural network training procedure. Experimental results show that our proposed method can improve inference accuracy by 10% similar to 30% under large variation, compared with uniform quantization method.
引用
收藏
页码:43 / 48
页数:6
相关论文
共 50 条
  • [31] ReRAM-Sharing: Fine-Grained Weight Sharing for ReRAM-Based Deep Neural Network Accelerator
    Song, Zhuoran
    Li, Dongyue
    He, Zhezhi
    Liang, Xiaoyao
    Jiang, Li
    2021 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2021,
  • [32] ReNEW: Enhancing Lifetime for ReRAM Crossbar based Neural Network Accelerators
    Wen, Wen
    Zhang, Youtao
    Yang, Jun
    2019 IEEE 37TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD 2019), 2019, : 487 - 496
  • [33] CNNWire: Boosting Convolutional Neural Network with Winograd on ReRAM based Accelerators
    Lin, Jilan
    Li, Shuangchen
    Hu, Xing
    Deng, Lei
    Xie, Yuan
    GLSVLSI '19 - PROCEEDINGS OF THE 2019 ON GREAT LAKES SYMPOSIUM ON VLSI, 2019, : 283 - 286
  • [34] REC: REtime Convolutional layers in energy harvesting ReRAM-based CNN accelerators
    Zhou, Kunyu
    Qiu, Keni
    PROCEEDINGS OF THE 19TH ACM INTERNATIONAL CONFERENCE ON COMPUTING FRONTIERS 2022 (CF 2022), 2022, : 185 - 188
  • [35] A Versatile ReRAM-based Accelerator for Convolutional Neural Networks
    Mao, Manqing
    Sun, Xiao Yu
    Peng, Xiaochen
    Yu, Shimeng
    Chakrabarti, Chaitali
    PROCEEDINGS OF THE 2018 IEEE INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING SYSTEMS (SIPS), 2018, : 211 - 216
  • [36] Fault-Free: A Framework for Analysis and Mitigation of Stuck-at-Fault on Realistic ReRAM-Based DNN Accelerators
    Shin, Hyein
    Kang, Myeonggu
    Kim, Lee-Sup
    IEEE TRANSACTIONS ON COMPUTERS, 2023, 72 (07) : 2011 - 2024
  • [37] A ReRAM-Based Convolutional Neural Network Accelerator Using the Analog Layer Normalization Technique
    Gi, Sang-Gyun
    Lee, Hyunkeun
    Jang, Jingon
    Lee, Byung-Geun
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2023, 70 (06) : 6442 - 6451
  • [38] An Energy-Efficient Inference Engine for a Configurable ReRAM-Based Neural Network Accelerator
    Zheng, Yang-Lin
    Yang, Wei-Yi
    Chen, Ya-Shu
    Han, Ding-Hung
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (03) : 740 - 753
  • [39] On-Line Fault Protection for ReRAM-Based Neural Networks
    Li, Wen
    Wang, Ying
    Liu, Cheng
    He, Yintao
    Liu, Lian
    Li, Huawei
    Li, Xiaowei
    IEEE TRANSACTIONS ON COMPUTERS, 2023, 72 (02) : 423 - 437
  • [40] PHANES: ReRAM-based Photonic Accelerator for Deep Neural Networks
    Liu, Yinyi
    Liu, Jiaqi
    Fu, Yuxiang
    Chen, Shixi
    Zhang, Jiaxu
    Xu, Jiang
    PROCEEDINGS OF THE 59TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC 2022, 2022, : 103 - 108