CipherFormer: Efficient Transformer Private Inference with Low Round Complexity

被引:0
|
作者
Wang, Weize [1 ]
Kuang, Yi [1 ]
机构
[1] Shanghai Jiao Tong Univ, Sch Cyber Sci & Engn, Shanghai, Peoples R China
来源
PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024 | 2024年
关键词
Private Inference; Transformer; Homomorphic Encryption; Garbled Circuit;
D O I
10.1109/CSCWD61410.2024.10580655
中图分类号
TP39 [计算机的应用];
学科分类号
081203 ; 0835 ;
摘要
There is a growing trend to outsource the inference task of large transformer models to cloud servers. However, this poses a severe threat to users' private data as they are exposed to cloud servers after uploading. Although several works attempted to provide private inference for transformer models, their hundreds of communication rounds limit the application scenarios. Motivated by the desire to minimize round complexity, we propose CipherFormer, a novel transformer private inference scheme using homomorphic encryption and garbled circuits. We present a protocol for quickly computing homomorphic matrix multiplications. We then modify the attention mechanism and design the corresponding garbled circuits. Furthermore, we show how to use a lightweight attention mechanism and mixed-bitwidth to reduce the inference latency while maintaining accuracy. In comparison with an advanced homomorphic encryption scheme on text classification tasks, our model improves accuracy by 3% to 11% while performing private inference with a 7.7x-11.9x speedup.
引用
收藏
页码:3054 / 3059
页数:6
相关论文
共 50 条
  • [1] Primer: Fast Private Transformer Inference on Encrypted Data
    Zheng, Mengxin
    Lou, Qian
    Jiang, Lei
    2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC, 2023,
  • [2] Constant-Round Private Function Evaluation with Linear Complexity
    Katz, Jonathan
    Malka, Lior
    ADVANCES IN CRYPTOLOGY - ASIACRYPT 2011, 2011, 7073 : 556 - +
  • [3] Selective Network Linearization for Efficient Private Inference
    Cho, Minsu
    Joshi, Ameya
    Garg, Siddharth
    Reagen, Brandon
    Hegde, Chinmay
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [4] Multi-tailed vision transformer for efficient inference
    Wang, Yunke
    Du, Bo
    Wang, Wenyuan
    Xu, Chang
    NEURAL NETWORKS, 2024, 174
  • [5] Efficient Transformer Inference with Statically Structured Sparse Attention
    Dai, Steve
    Genc, Hasan
    Venkatesan, Rangharajan
    Khailany, Brucek
    2023 60TH ACM/IEEE DESIGN AUTOMATION CONFERENCE, DAC, 2023,
  • [6] DeepSpeed-Inference: Enabling Efficient Inference of Transformer Models at Unprecedented Scale
    Aminabadi, Reza Yazdani
    Rajbhandari, Samyam
    Awan, Ammar Ahmad
    Li, Cheng
    Li, Du
    Zheng, Elton
    Ruwase, Olatunji
    Smith, Shaden
    Zhang, Minjia
    Rasley, Jeff
    He, Yuxiong
    SC22: INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, 2022,
  • [7] SecureTLM: Private inference for transformer-based large model with MPC
    Chen, Yuntian
    Meng, Xianjia
    Shi, Zhiying
    Ning, Zhiyuan
    Lin, Jingzhi
    INFORMATION SCIENCES, 2024, 667
  • [8] Efficient Bayesian inference using adversarial machine learning and low-complexity surrogate models
    Na, Jonggeol
    Bak, Ji Hyun
    Sahinidis, Nikolaos V.
    COMPUTERS & CHEMICAL ENGINEERING, 2021, 151
  • [9] Learning Bayesian networks with low inference complexity
    Benjumeda M.
    Larrañaga P.
    Bielza C.
    Progress in Artificial Intelligence, 2016, 5 (1) : 15 - 26
  • [10] SILT: Efficient transformer training for inter-lingual inference
    Huertas-Tato, Javier
    Martin, Alejandro
    Camacho, David
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 200