CipherFormer: Efficient Transformer Private Inference with Low Round Complexity

被引:0
|
作者
Wang, Weize [1 ]
Kuang, Yi [1 ]
机构
[1] Shanghai Jiao Tong Univ, Sch Cyber Sci & Engn, Shanghai, Peoples R China
来源
PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024 | 2024年
关键词
Private Inference; Transformer; Homomorphic Encryption; Garbled Circuit;
D O I
10.1109/CSCWD61410.2024.10580655
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
There is a growing trend to outsource the inference task of large transformer models to cloud servers. However, this poses a severe threat to users' private data as they are exposed to cloud servers after uploading. Although several works attempted to provide private inference for transformer models, their hundreds of communication rounds limit the application scenarios. Motivated by the desire to minimize round complexity, we propose CipherFormer, a novel transformer private inference scheme using homomorphic encryption and garbled circuits. We present a protocol for quickly computing homomorphic matrix multiplications. We then modify the attention mechanism and design the corresponding garbled circuits. Furthermore, we show how to use a lightweight attention mechanism and mixed-bitwidth to reduce the inference latency while maintaining accuracy. In comparison with an advanced homomorphic encryption scheme on text classification tasks, our model improves accuracy by 3% to 11% while performing private inference with a 7.7x-11.9x speedup.
引用
收藏
页码:3054 / 3059
页数:6
相关论文
共 50 条
  • [21] PolyKervNets: Activation-free Neural Networks For Efficient Private Inference
    Aremu, Toluwani
    Nandakumar, Karthik
    2023 IEEE CONFERENCE ON SECURE AND TRUSTWORTHY MACHINE LEARNING, SATML, 2023, : 593 - 604
  • [22] Private Inference for Deep Neural Networks: A Secure, Adaptive, and Efficient Realization
    Cheng, Ke
    Xi, Ning
    Liu, Ximeng
    Zhu, Xinghui
    Gao, Haichang
    Zhang, Zhiwei
    Shen, Yulong
    IEEE TRANSACTIONS ON COMPUTERS, 2023, 72 (12) : 3519 - 3531
  • [23] LCFormer: linear complexity transformer for efficient image super-resolution
    Gao, Xiang
    Wu, Sining
    Zhou, Ying
    Wang, Fan
    Hu, Xiaopeng
    MULTIMEDIA SYSTEMS, 2024, 30 (04)
  • [24] Low-randomness constant-round private XOR computations
    Blundo, Carlo
    Galdi, Clemente
    Persiano, Giuseppe
    INTERNATIONAL JOURNAL OF INFORMATION SECURITY, 2007, 6 (01) : 15 - 26
  • [25] Co-Designing Transformer Architectures for Distributed Inference With Low Communication
    Du, Jiangsu
    Wei, Yuanxin
    Ye, Shengyuan
    Jiang, Jiazhi
    Chen, Xu
    Huang, Dan
    Lu, Yutong
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2025, 36 (04) : 717 - 730
  • [26] Round-Efficient Private Stable Matching from Additive Homomorphic Encryption
    Teruya, Tadanori
    Sakuma, Jun
    INFORMATION SECURITY (ISC 2013), 2015, 7807 : 69 - 86
  • [27] Low-randomness constant-round private XOR computations
    Carlo Blundo
    Clemente Galdi
    Giuseppe Persiano
    International Journal of Information Security, 2007, 6 : 15 - 26
  • [28] An Efficient Piecewise Linear Approximation of Non-linear Operations for Transformer Inference
    Lu, Haodong
    Mei, Qichang
    Wang, Kun
    2023 IEEE 31ST ANNUAL INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE CUSTOM COMPUTING MACHINES, FCCM, 2023, : 206 - 206
  • [29] Input Compression with Positional Consistency for Efficient Training and Inference of Transformer Neural Networks
    Nagarajan, Amrit
    Raghunathan, Anand
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES: RESEARCH TRACK, PT V, ECML PKDD 2024, 2024, 14945 : 73 - 88
  • [30] Energy-Efficient Online Scheduling of Transformer Inference Services on GPU Servers
    Wang, Yuxin
    Wang, Qiang
    Chu, Xiaowen
    IEEE TRANSACTIONS ON GREEN COMMUNICATIONS AND NETWORKING, 2022, 6 (03): : 1649 - 1659