A Quantized Training Method to Enhance Accuracy of ReRAM-based Neuromorphic Systems

被引:7
|
作者
Yang, Qing [1 ]
Li, Hai [1 ]
Wu, Qing [2 ]
机构
[1] Duke Univ, Dept Elect & Comp Engn, Durham, NC 27708 USA
[2] US Air Force, Res Lab, Informat Directorate, New York, NY USA
基金
美国国家科学基金会;
关键词
D O I
10.1109/ISCAS.2018.8351327
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Deep neural networks (DNNs) are tremendously applied in artificial intelligence field. While the performance of DNNs is continuously improved by more complicated and deeper structures, the feasibility of deployment on embedded system remains as a critical problem. Neuromorphic system designs based on resistive random-access memory (ReRAM) provide an opportunity for power-efficient DNN employment. However, it encounters the challenge of limited programming resolution. A quantized training method is proposed in this paper to enhance the performance of neuromorphic systems based on ReRAM. Different from previous methods in which a dedicated regularization term is used in loss function to constrain parameter distribution, our quantized training method deals with training and quantization at the same time to alleviate the impact of limited parameter precision. Models with discrete parameters obtained after training can be directly mapped onto ReRAM devices. We implement experiments on image recognition tasks using multi-layer perceptron (MLP) and convolution neural network (CNN). The results verify that the quantized training method can approximate the accuracy of full-precision training, e.g., a two-layer MLP based on binary ReRAM only decreases the classification accuracy by 0.25% for MNIST dataset. In addition, we carefully investigate and present the importance of layer size under the ReRAM's low programming resolution, the different parameter resolution demands for convolution layer and fully connected layer, and system robustness to ReRAM variations after quantized training. The codes are available at https://github.com/qingyangqing/quantized-rram-net.
引用
收藏
页数:5
相关论文
共 50 条
  • [21] Learning to Predict IR Drop with Effective Training for ReRAM-based Neural Network Hardware
    Lee, Sugil
    Jung, Giju
    Fouda, Mohammed E.
    Lee, Jongeun
    Eltawil, Ahmed
    Kurdahi, Fadi
    PROCEEDINGS OF THE 2020 57TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2020,
  • [22] A Unified Framework for Training, Mapping and Simulation of ReRAM-Based Convolutional Neural Network Acceleration
    Liu, He
    Han, Jianhui
    Zhang, Youhui
    IEEE COMPUTER ARCHITECTURE LETTERS, 2019, 18 (01) : 63 - 66
  • [23] ESSENCE: Exploiting Structured Stochastic Gradient Pruning for Endurance-Aware ReRAM-Based In-Memory Training Systems
    Yang, Xiaoxuan
    Yang, Huanrui
    Doppa, Janardhan Rao
    Pande, Partha Pratim
    Chakrabartys, Krishnendu
    Li, Hai
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (07) : 2187 - 2199
  • [24] Energy-Efficient ReRAM-based ML Training via Mixed Pruning and Reconfigurable ADC
    Ogbogu, Chukwufumnanya
    Soumen, Mohapatra
    Joardar, Biresh Kumar
    Doppa, Janardhan Rao
    Heo, Deuk
    Chakrabarty, Krishnendu
    Pande, Partha Pratim
    2023 IEEE/ACM INTERNATIONAL SYMPOSIUM ON LOW POWER ELECTRONICS AND DESIGN, ISLPED, 2023,
  • [25] Offline Training-Based Mitigation of IR Drop for ReRAM-Based Deep Neural Network Accelerators
    Lee, Sugil
    Fouda, Mohammed E.
    Lee, Jongeun
    Eltawil, Ahmed M.
    Kurdahi, Fadi
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (02) : 521 - 532
  • [26] Training-Free Stuck-At Fault Mitigation for ReRAM-Based Deep Learning Accelerators
    Quan, Chenghao
    Fouda, Mohammed E.
    Lee, Sugil
    Jung, Giju
    Lee, Jongeun
    Eltawil, Ahmed E.
    Kurdahi, Fadi
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (07) : 2174 - 2186
  • [27] Online Fault Detection in ReRAM-Based Computing Systems by Monitoring Dynamic Power Consumption
    Liu, Mengyun
    Chakrabarty, Krishnendu
    2020 IEEE INTERNATIONAL TEST CONFERENCE (ITC), 2020,
  • [28] A Resource-Efficient Weight Quantization and Mapping Method for Crossbar Arrays in ReRAM-based Computing-in-Memory Systems
    Ma, Mingyuan
    Jiang, Wei
    Liu, Juntao
    Dui, Li
    Ma, Zhongyuan
    Du, Yuan
    2024 IEEE THE 20TH ASIA PACIFIC CONFERENCE ON CIRCUITS AND SYSTEMS, APCCAS 2024, 2024, : 169 - 173
  • [29] Accelerating Graph Neural Network Training on ReRAM-Based PIM Architectures via Graph and Model Pruning
    Ogbogu, Chukwufumnanya O.
    Arka, Aqeeb Iqbal
    Pfromm, Lukas
    Joardar, Biresh Kumar
    Doppa, Janardhan Rao
    Chakrabarty, Krishnendu
    Pande, Partha Pratim
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (08) : 2703 - 2716
  • [30] High-Throughput Training of Deep CNNs on ReRAM-Based Heterogeneous Architectures via Optimized Normalization Layers
    Joardar, Biresh Kumar
    Deshwal, Aryan
    Doppa, Janardhan Rao
    Pande, Partha Pratim
    Chakrabarty, Krishnendu
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2022, 41 (05) : 1537 - 1549