GALA: Greedy ComputAtion for Linear Algebra in Privacy-Preserved Neural Networks

被引:10
|
作者
Zhang, Qiao [1 ]
Xin, Chunsheng [1 ]
Wu, Hongyi [1 ]
机构
[1] Old Dominion Univ, Norfolk, VA 23529 USA
基金
美国国家科学基金会;
关键词
MACHINE; ATTACKS;
D O I
10.14722/ndss.2021.24351
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Machine Learning as a Service (MLaaS) is enabling a wide range of smart applications on end devices. However, privacy still remains a fundamental challenge. The schemes that exploit Homomorphic Encryption (HE)-based linear computations and Garbled Circuit (GC)-based nonlinear computations have demonstrated superior performance to enable privacy-preserved MLaaS. Nevertheless, there is still a significant gap in the computation speed. Our investigation has found that the HE-based linear computation dominates the total computation time for state-of-the-art deep neural networks. Furthermore, the most time-consuming component of the HE-based linear computation is a series of Permutation (Perm) operations that are imperative for dot product and convolution in privacy-preserved MLaaS. This work focuses on a deep optimization of the HEbased linear computations to minimize the Perm operations, thus substantially reducing the overall computation time. To this end, we propose GALA: Greedy computAtion for Linear Algebra in privacy-preserved neural networks, which views the HE-based linear computation as a series of Homomorphic Add, Mult and Perm operations and chooses the least expensive operation in each linear computation step to reduce the overall cost. GALA makes the following contributions: (1) It introduces a row-wise weight matrix encoding and combines the share generation that is needed for the GC-based nonlinear computation, to reduce the Perm operations for the dot product; (2) It designs a first-Add-second-Perm approach (named kernel grouping) to reduce Perm operations for convolution. As such, GALA efficiently reduces the cost for the HE-based linear computation, which is a critical building block in almost all of the recent frameworks for privacy-preserved neural networks, including GAZELLE (Usenix Security'18), DELPHI (Usenix Security'20), and CrypTFlow2 (CCS'20). With its deep optimization of the HE-based linear computation, GALA can be a plug-and-play module integrated into these systems to further boost their efficiency. Our experiments show that it achieves a significant speedup up to 700x for the dot product and 14x for the convolution computation under different data dimensions. Meanwhile, GALA demonstrates an encouraging runtime boost by 2.5x , 2.7x , 3.2x , 83x , 7.7x , and 7.5x over GAZELLE and 6.5x , 6x , 5.7x , 4.5x , 4.2x , and 4.1 x over CrypTFlow2, on AlexNet, VGG, ResNet-18, ResNet-50, ResNet-101, and ResNet-152, respectively.
引用
收藏
页数:16
相关论文
共 26 条
  • [21] Secure multi-party computation in deep learning : Enhancing privacy in distributed neural networks
    Sagar, P. Vidya
    Ghanimi, Hayder M. A.
    Prabhu, L. Arokia Jesu
    Raja, L.
    Dadheech, Pankaj
    Sengan, Sudhakar
    JOURNAL OF DISCRETE MATHEMATICAL SCIENCES & CRYPTOGRAPHY, 2024, 27 (2A): : 249 - 259
  • [22] Find and Dig: A Privacy-Preserving Image Processing Mechanism in Deep Neural Networks for Mobile Computation
    Huang, Hongyu
    Zhao, Hongyuan
    Hu, Chunqiang
    Chen, Chao
    Li, Yantao
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [23] Mixing neural networks, continuation and symbolic computation to solve parametric systems of non linear equations
    Merlet J.-P.
    Neural Networks, 2024, 176
  • [24] Machine Learning-Aided Numerical Linear Algebra: Convolutional Neural Networks for the Efficient Preconditioner Generation
    Goetz, Markus
    Anzt, Hartwig
    SCALA 2018: PROCEEDINGS OF 2018 IEEE/ACM 9TH WORKSHOP ON LATEST ADVANCES IN SCALABLE ALGORITHMS FOR LARGE-SCALE SYSTEMS (SCALA), 2018, : 49 - 56
  • [25] Gradient Computation of Continuous-Time Cellular Neural/Nonlinear Networks with Linear Templates via the CNN Universal Machine
    Mátyás Brendel
    Tamás Roska
    Gusztáv Bártfai
    Neural Processing Letters, 2002, 16 : 111 - 120
  • [26] Gradient computation of continuous-time cellular neural/nonlinear networks with linear templates via the CNN universal machine
    Brendel, M
    Roska, T
    Bártfai, G
    NEURAL PROCESSING LETTERS, 2002, 16 (02) : 111 - 120