SEAR: Secure and Efficient Aggregation for Byzantine-Robust Federated Learning

被引:42
|
作者
Zhao, Lingchen [1 ]
Jiang, Jianlin [2 ]
Feng, Bo [3 ]
Wang, Qian [1 ]
Shen, Chao [4 ]
Li, Qi [5 ,6 ]
机构
[1] Wuhan Univ, Sch Cyber Sci & Engn, Key Lab Aerosp Informat Secur & Trusted Comp, Minist Educ, Wuhan 430072, Hubei, Peoples R China
[2] Wuhan Univ, Sch Comp Sci, Wuhan 430072, Hubei, Peoples R China
[3] Northeastern Univ, Khoury Coll Comp Sci, Boston, MA 02115 USA
[4] Xi An Jiao Tong Univ, Sch Cyber Sci & Engn, MOE Key Lab Intelligent Networks & Network Secur, Xian 710049, Shaanxi, Peoples R China
[5] Tsinghua Univ, Inst Network Sci & Cyberspace, Beijing 100084, Peoples R China
[6] Tsinghua Univ, Beijing Natl Res Ctr Informat Sci & Technol BNRis, Beijing 100084, Peoples R China
基金
国家重点研发计划;
关键词
Servers; Computational modeling; Collaborative work; Data models; Privacy; Cryptography; Training data; Federated learning; secure aggregation; trusted execution environment;
D O I
10.1109/TDSC.2021.3093711
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning facilitates the collaborative training of a global model among distributed clients without sharing their training data. Secure aggregation, a new security primitive for federated learning, aims to preserve the confidentiality of both local models and training data. Unfortunately, existing secure aggregation solutions fail to defend against Byzantine failures that are common in distributed computing systems. In this work, we propose a new secure and efficient aggregation framework, SEAR, for Byzantine-robust federated learning. Relying on the trusted execution environment, i.e., Intel SGX, SEAR protects clients' private models while enabling Byzantine resilience. Considering the limitation of the current Intel SGX's architecture (i.e., the limited trusted memory), we propose two data storage modes to efficiently implement aggregation algorithms efficiently in SGX. Moreover, to balance the efficiency and performance of aggregation, we propose a sampling-based method to efficiently detect Byzantine failures without degrading the global model's performance. We implement and evaluate SEAR in a LAN environment, and the experiment results show that SEAR is computationally efficient and robust to Byzantine adversaries. Compared to the previous practical secure aggregation framework, SEAR improves aggregation efficiency by 4-6 times while supporting Byzantine resilience at the same time.
引用
收藏
页码:3329 / 3342
页数:14
相关论文
共 50 条
  • [21] Byzantine-Robust Federated Linear Bandits
    Jadbabaie, Ali
    Li, Haochuan
    Qian, Jian
    Tian, Yi
    [J]. 2022 IEEE 61ST CONFERENCE ON DECISION AND CONTROL (CDC), 2022, : 5206 - 5213
  • [22] FedCom: Byzantine-Robust Federated Learning Using Data Commitment
    Zhao, Bo
    Wang, Tao
    Fang, Liming
    [J]. ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 33 - 38
  • [23] Communication-Efficient and Byzantine-Robust Federated Learning for Mobile Edge Computing Networks
    Zhang, Zhuangzhuang
    Wl, Libing
    He, Debiao
    Li, Jianxin
    Cao, Shuqin
    Wu, Xianfeng
    [J]. IEEE NETWORK, 2023, 37 (04): : 112 - 119
  • [24] Local Model Poisoning Attacks to Byzantine-Robust Federated Learning
    Fang, Minghong
    Cao, Xiaoyu
    Jia, Jinyuan
    Gong, Neil Nenqiang
    [J]. PROCEEDINGS OF THE 29TH USENIX SECURITY SYMPOSIUM, 2020, : 1623 - 1640
  • [25] SIREN: Byzantine-robust Federated Learning via Proactive Alarming
    Guo, Hanxi
    Wang, Hao
    Song, Tao
    Hua, Yang
    Lv, Zhangcheng
    Jin, Xiulang
    Xue, Zhengui
    Ma, Ruhui
    Guan, Haibing
    [J]. PROCEEDINGS OF THE 2021 ACM SYMPOSIUM ON CLOUD COMPUTING (SOCC '21), 2021, : 47 - 60
  • [26] FLTrust: Byzantine-robust Federated Learning via Trust Bootstrapping
    Cao, Xiaoyu
    Fang, Minghong
    Liu, Jia
    Gong, Neil Zhenqiang
    [J]. 28TH ANNUAL NETWORK AND DISTRIBUTED SYSTEM SECURITY SYMPOSIUM (NDSS 2021), 2021,
  • [27] FLForest: Byzantine-robust Federated Learning through Isolated Forest
    Wang, Tao
    Zhao, Bo
    Fang, Liming
    [J]. 2022 IEEE 28TH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS, ICPADS, 2022, : 296 - 303
  • [28] Lightweight Byzantine-Robust and Privacy-Preserving Federated Learning
    Lu, Zhi
    Lu, Songfeng
    Cui, Yongquan
    Wu, Junjun
    Nie, Hewang
    Xiao, Jue
    Yi, Zepu
    [J]. EURO-PAR 2024: PARALLEL PROCESSING, PART II, EURO-PAR 2024, 2024, 14802 : 274 - 287
  • [29] Byzantine-Robust Federated Learning with Variance Reduction and Differential Privacy
    Zhang, Zikai
    Hu, Rui
    [J]. 2023 IEEE CONFERENCE ON COMMUNICATIONS AND NETWORK SECURITY, CNS, 2023,
  • [30] BSR-FL: An Efficient Byzantine-Robust Privacy-Preserving Federated Learning Framework
    Zeng, Honghong
    Li, Jie
    Lou, Jiong
    Yuan, Shijing
    Wu, Chentao
    Zhao, Wei
    Wu, Sijin
    Wang, Zhiwen
    [J]. IEEE TRANSACTIONS ON COMPUTERS, 2024, 73 (08) : 2096 - 2110