An Empirical Fault Vulnerability Exploration of ReRAM-Based Process-in-Memory CNN Accelerators

被引:0
|
作者
Dorostkar, Aniseh [1 ]
Farbeh, Hamed [1 ]
Zarandi, Hamid R. [1 ]
机构
[1] Amirkabir Univ Technol, Tehran Polytech, Tehran 158754413, Iran
关键词
Circuit faults; Neural networks; Resistance; Random access memory; Virtual machine monitors; Matrix converters; Kernel; Convolutional neural networks (CNNs); fault vulnerability; hardware accelerators; processing-in-memory (PIM); resistive random-access memory (ReRAM); RRAM DEVICES;
D O I
10.1109/TR.2024.3405825
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Resistive random-access memory (ReRAM)-based processing-in-memory (PIM) accelerator is a promising platform for processing massively memory intensive matrix-vector multiplications of neural networks in parallel domain, due to its capability of analog computation, ultra-high density, near-zero leakage current, and nonvolatility. Despite many advantages, ReRAM-based accelerators are highly error-prone due to limitations of technology fabrication that lead to process variations and defects. These limitations degrade the accuracy of deep convolutional neural networks (CNNs) (Deep CNNs) running on PIM accelerators. While these CNNs accelerators are widely deployed in safety-critical systems, their vulnerability to fault is not well explored. In this article, we have developed a fault-injection framework to investigate the vulnerability of large-scale CNNs at both software- and hardware-level of inference phases. Faulty ReRAM devices are another reliability challenges due to significant degradation of classification accuracy when CNN parameters are mapped to the accelerators. To investigate this challenge, we map the CNN learning parameter to the ReRAM crossbar and inject faults into crossbar arrays. The proposed framework analyzes the impact of stuck-at high (SaH) and stuck-at low (SaL) fault models on different layers and locations of CNN learning parameters. By performing extensive fault injections, we illustrate that the vulnerability behavior of ReRAM-based PIM accelerator for CNNs is greatly impressible to the types and depth of layers, the location of the learning parameter in every layer, and the value and types of faults. Our observations show that different models have different vulnerabilities to faults. Specifically, we show that SaL further reduces classification accuracy than SaH.
引用
收藏
页码:1 / 15
页数:15
相关论文
共 50 条
  • [1] Mathematical Framework for Optimizing Crossbar Allocation for ReRAM-based CNN Accelerators
    Li, Wanqian
    Han, Yinhe
    Chen, Xiaoming
    ACM TRANSACTIONS ON DESIGN AUTOMATION OF ELECTRONIC SYSTEMS, 2024, 29 (01)
  • [2] REC: REtime Convolutional layers in energy harvesting ReRAM-based CNN accelerators
    Zhou, Kunyu
    Qiu, Keni
    PROCEEDINGS OF THE 19TH ACM INTERNATIONAL CONFERENCE ON COMPUTING FRONTIERS 2022 (CF 2022), 2022, : 185 - 188
  • [3] Hardware attacks on ReRAM-based AI accelerators
    Heidary, Masoud
    Joardar, Biresh Kumar
    17TH IEEE DALLAS CIRCUITS AND SYSTEMS CONFERENCE, DCAS 2024, 2024,
  • [4] FARe: Fault-Aware GNN Training on ReRAM-based PIM Accelerators
    Dhingra, Pratyush
    Ogbogu, Chukwufumnanya
    Joardar, Biresh Kumar
    Doppa, Janardhan Rao
    Kalyanaraman, Ananth
    Pande, Partha Pratim
    2024 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION, DATE, 2024,
  • [5] DL-RSIM: A Reliability and Deployment Strategy Simulation Framework for ReRAM-based CNN Accelerators
    Lin, Wei-Ting
    Cheng, Hsiang-Yun
    Yang, Chia-Lin
    Lin, Meng-Yao
    Lien, Kai
    Hu, Han-Wen
    Chang, Hung-Sheng
    Li, Hsiang-Pang
    Chang, Meng-Fan
    Tsou, Yen-Ting
    Nien, Chin-Fu
    ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2022, 21 (03)
  • [6] REC: REtime Convolutional Layers to Fully Exploit Harvested Energy for ReRAM-based CNN Accelerators
    Zhou, Kunyu
    Qiu, Keni
    ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2024, 23 (06) : 33 - 33
  • [7] MaxTracker: Continuously Tracking the Maximum Computation Progress for Energy Harvesting ReRAM-based CNN Accelerators
    Qiu, Keni
    Jao, Nicholas
    Zhou, Kunyu
    Liu, Yongpan
    Sampson, Jack
    Kandemir, Mahmut Taylan
    Narayanan, Vijaykrishnan
    ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2021, 20 (05)
  • [8] Training-Free Stuck-At Fault Mitigation for ReRAM-Based Deep Learning Accelerators
    Quan, Chenghao
    Fouda, Mohammed E.
    Lee, Sugil
    Jung, Giju
    Lee, Jongeun
    Eltawil, Ahmed E.
    Kurdahi, Fadi
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (07) : 2174 - 2186
  • [9] APQ: Automated DNN Pruning and Quantization for ReRAM-Based Accelerators
    Yang, Siling
    He, Shuibing
    Duan, Hexiao
    Chen, Weijian
    Zhang, Xuechen
    Wu, Tong
    Yin, Yanlong
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2023, 34 (09) : 2498 - 2511
  • [10] Mixed Precision Quantization for ReRAM-based DNN Inference Accelerators
    Huang, Sitao
    Ankit, Aayush
    Silveira, Plinio
    Antunes, Rodrigo
    Chalamalasetti, Sai Rahul
    El Hajj, Izzat
    Kim, Dong Eun
    Aguiar, Glaucimar
    Bruel, Pedro
    Serebryakov, Sergey
    Xu, Cong
    Li, Can
    Faraboschi, Paolo
    Strachan, John Paul
    Chen, Deming
    Roy, Kaushik
    Hwu, Wen-mei
    Milojicic, Dejan
    2021 26TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2021, : 372 - 377