Towards Deep Neural Network Training on Encrypted Data

被引:67
|
作者
Nandakumar, Karthik [1 ]
Ratha, Nalini [2 ]
Pankanti, Sharath [2 ]
Halevi, Shai [3 ]
机构
[1] IBM Res, Singapore, Singapore
[2] IBM Res, Yorktown Hts, NY 10598 USA
[3] Algorand Fdn, Boston, MA USA
关键词
FULLY HOMOMORPHIC ENCRYPTION;
D O I
10.1109/CVPRW.2019.00011
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While deep learning is a valuable tool for solving many tough problems in computer vision., the success of deep learning models is typically determined by: (1) availability of sufficient training data, (ii) access to extensive computational resources, and (iii) expertise in selecting the right model and hyperparameters for the selected task. Often, the availability of data is the hard part due to compliance, legal, and privacy constraints. Cryptographic techniques such as fully homomorphic encryption, (THE) offer a potential solution by enabling processing on encrypted data. While prior work has been done on using FHE for inferencing, training a deep neural network in the encrypted domain is an extremely challenging task due to the computational complexity of the operations involved. In this paper, we evaluate the feasibility of training neural networks on encrypted data in a completely non-interactive way. Our proposed system uses the open-source FHE toolkit HElib to implement a Stochastic Gradient Descent (SGD)-based training of a neural network. We show that encrypted training can be made more computationally efficient by (i) simplifyinj_s, the network with minimal degradation of accuracy, choosing appropriate data representation and resolution, and (iii) packing the data elements within the ciphertext in a smart way so as to minimize the number of operations and facilitate parallelization of FHE computations. Based on the above optimizations, we demonstrate that it is possible to achieve more than 50x speed up while training a fully-connected neural network on the MNIST dataset while achieving reasonable accuracy (96%). Though the cost of training a complex deep learning model from scratch on encrypted data is still very high, this work establishes a solid baseline and paves the way for relatively simpler tasks such as fine-tuning of deep learning models based on encrypted data to be implemented in the near future.
引用
收藏
页码:40 / 48
页数:9
相关论文
共 50 条
  • [1] Deep Neural Network Model over Encrypted Data
    Li, Weixun
    Sun, Guanghui
    Wang, Yajun
    Yuan, Long
    Gao, Minghui
    Dong, Yan
    Wang, Chen
    [J]. EMERGING INFORMATION SECURITY AND APPLICATIONS, EISA 2023, 2024, 2004 : 107 - 117
  • [2] SHE: A Fast and Accurate Deep Neural Network for Encrypted Data
    Lou, Qian
    Jiang, Lei
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [3] Glyph: Fast and Accurately Training Deep Neural Networks on Encrypted Data
    Lou, Qian
    Feng, Bo
    Fox, Geoffrey C.
    Jiang, Lei
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [4] Deep Binarized Convolutional Neural Network Inferences over Encrypted Data
    Zhou, Junwei
    Li, Junjiong
    Panaousis, Emmanouil
    Liang, Kaitai
    [J]. 2020 7TH IEEE INTERNATIONAL CONFERENCE ON CYBER SECURITY AND CLOUD COMPUTING (CSCLOUD 2020)/2020 6TH IEEE INTERNATIONAL CONFERENCE ON EDGE COMPUTING AND SCALABLE CLOUD (EDGECOM 2020), 2020, : 160 - 167
  • [5] Automatic Generation of Work Support Behavior with Smart Glasses based on the Deep Neural Network Corresponding to Encrypted Training Data
    Hashimoto, Kohjiro
    Miyosawa, Tadashi
    Yamada, Tetsuyasu
    [J]. IECON 2021 - 47TH ANNUAL CONFERENCE OF THE IEEE INDUSTRIAL ELECTRONICS SOCIETY, 2021,
  • [6] Accelerating Data Loading in Deep Neural Network Training
    Yang, Chih-Chieh
    Cong, Guojing
    [J]. 2019 IEEE 26TH INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE COMPUTING, DATA, AND ANALYTICS (HIPC), 2019, : 235 - 245
  • [7] CryptoNN: Training Neural Networks over Encrypted Data
    Xu, Runhua
    Joshi, James B. D.
    Li, Chao
    [J]. 2019 39TH IEEE INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS (ICDCS 2019), 2019, : 1199 - 1209
  • [8] Trompt: Towards a Better Deep Neural Network for Tabular Data
    Chen, Kuan-Yu
    Chiang, Ping-Han
    Chou, Hsin-Rung
    Chen, Ting-Wei
    Chang, Darby Tien-Hao
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202
  • [9] Gist: Efficient Data Encoding for Deep Neural Network Training
    Jain, Animesh
    Phanishayee, Amar
    Mars, Jason
    Tang, Lingjia
    Pekhimenko, Gennady
    [J]. 2018 ACM/IEEE 45TH ANNUAL INTERNATIONAL SYMPOSIUM ON COMPUTER ARCHITECTURE (ISCA), 2018, : 776 - 789
  • [10] Deep Neural Networks Classification over Encrypted Data
    Hesamifard, Ehsan
    Takabi, Hassan
    Ghasemi, Mehdi
    [J]. PROCEEDINGS OF THE NINTH ACM CONFERENCE ON DATA AND APPLICATION SECURITY AND PRIVACY (CODASPY '19), 2019, : 97 - 108