Enabling Secure NVM-Based in-Memory Neural Network Computing by Sparse Fast Gradient Encryption

被引:16
|
作者
Cai, Yi [1 ]
Chen, Xiaoming [2 ]
Tian, Lu [3 ]
Wang, Yu [1 ]
Yang, Huazhong [1 ]
机构
[1] Tsinghua Univ, Beijing Natl Res Ctr Informat Sci & Technol BNRis, Dept Elect Engn, Beijing 100084, Peoples R China
[2] Chinese Acad Sci, Inst Comp Technol, State Key Lab Comp Architecture, Beijing 100864, Peoples R China
[3] Xilinx Inc, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Artificial neural networks; Nonvolatile memory; Encryption; Computational modeling; Hardware; Non-volatile memory (NVM); compute-in-memory (CIM); neural network; security; encryption; ATTACKS;
D O I
10.1109/TC.2020.3017870
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Neural network (NN) computing is energy-consuming on traditional computing systems, owing to the inherent memory wall bottleneck of the von Neumann architecture and the Moore's Law being approaching the end. Non-volatile memories (NVMs) have been demonstrated as promising alternatives for constructing computing-in-memory (CIM) systems to accelerate NN computing. However, NVM-based NN computing systems are vulnerable to the confidentiality attacks because the weight parameters persist in memory when the system is powered off, enabling an adversary with physical access to extract the well-trained NN models. The goal of this article is to find a solution for thwarting the confidentiality attacks. We define and model the weight encryption problem. Then we propose an effective framework, containing a sparse fast gradient encryption (SFGE) method and a runtime encryption scheduling (RES) scheme, to guarantee the confidentiality security of NN models with a negligible performance overhead. Moreover, we improve the SFGE method by incrementally generating the encryption keys. Additionally, we provide variants of the encryption method to better fit quantized models and various mapping strategies. The experiments demonstrate that only encrypting an extremely small proportion of the weights (e.g., 20 weights per layer in ResNet-101), the NN models can be strictly protected.
引用
收藏
页码:1596 / 1610
页数:15
相关论文
共 50 条
  • [31] Gradient descent-based programming of analog in-memory computing cores
    Buchel, J.
    Vasilopoulos, A.
    Kersting, B.
    Odermatt, F.
    Brew, K.
    Ok, I.
    Choi, S.
    Saraf, I.
    Chan, V.
    Philip, T.
    Saulnier, N.
    Narayanan, V.
    Le Gallo, M.
    Sebastian, A.
    2022 INTERNATIONAL ELECTRON DEVICES MEETING, IEDM, 2022,
  • [32] SDESA: secure cloud computing with gradient deep belief network and congruential advanced encryption
    Rani, Sheeja
    Raj, P. V. Pravija
    Khedr, Ahmed M.
    JOURNAL OF SUPERCOMPUTING, 2024, 80 (15): : 23147 - 23176
  • [33] Sparse and Robust RRAM-based Efficient In-memory Computing for DNN Inference
    Meng, Jian
    Yeo, Injune
    Shim, Wonbo
    Yang, Li
    Fan, Deliang
    Yu, Shimeng
    Seo, Jae-Sun
    2022 IEEE INTERNATIONAL RELIABILITY PHYSICS SYMPOSIUM (IRPS), 2022,
  • [34] Sparse and Robust RRAM-based Efficient In-memory Computing for DNN Inference
    Meng, Jian
    Yeo, Injune
    Yang, Li
    Fan, Deliang
    Seo, Jae-sun
    Yu, Shimeng
    Shim, Wonbo
    2022 IEEE INTERNATIONAL RELIABILITY PHYSICS SYMPOSIUM (IRPS), 2022,
  • [35] A large-scale in-memory computing for deep neural network with trained quantization
    Cheng, Yuan
    Wang, Chao
    Chen, Hai-Bao
    Yu, Hao
    INTEGRATION-THE VLSI JOURNAL, 2019, 69 : 345 - 355
  • [36] AnalogNAS: A Neural Network Design Framework for Accurate Inference with Analog In-Memory Computing
    Benmeziane, Hadjer
    Lammie, Corey
    Boybat, Irem
    Rasch, Malte
    Le Gallo, Manuel
    Tsai, Hsinyu
    Muralidhar, Ramachandran
    Niar, Smail
    Hamza, Ouarnoughi
    Narayanan, Vijay
    Sebastian, Abu
    El Maghraoui, Kaoutar
    2023 IEEE INTERNATIONAL CONFERENCE ON EDGE COMPUTING AND COMMUNICATIONS, EDGE, 2023, : 233 - 244
  • [37] Secure Convolution Neural Network Inference Based on Homomorphic Encryption
    Song, Chen
    Huang, Ruwei
    APPLIED SCIENCES-BASEL, 2023, 13 (10):
  • [38] A Multilevel Cell STT-MRAM-Based Computing In-Memory Accelerator for Binary Convolutional Neural Network
    Pan, Yu
    Ouyang, Peng
    Zhao, Yinglin
    Kang, Wang
    Yin, Shouyi
    Zhang, Youguang
    Zhao, Weisheng
    Wei, Shaojun
    IEEE TRANSACTIONS ON MAGNETICS, 2018, 54 (11)
  • [39] A MLC STT-MRAM based Computing in-Memory Architec-ture for Binary Neural Network
    Pan, Y.
    Ouyang, P.
    Zhao, Y.
    Kang, W.
    Yin, S.
    Zhang, Y.
    Zhao, W.
    Wei, S.
    2018 IEEE INTERNATIONAL MAGNETIC CONFERENCE (INTERMAG), 2018,
  • [40] OxRRAM-Based Analog in-Memory Computing for Deep Neural Network Inference: A Conductance Variability Study
    Doevenspeck, J.
    Degraeve, R.
    Fantini, A.
    Cosemans, S.
    Mallik, A.
    Debacker, P.
    Verkest, D.
    Lauwereins, R.
    Dehaene, W.
    IEEE TRANSACTIONS ON ELECTRON DEVICES, 2021, 68 (05) : 2301 - 2305