Communication-Efficient and Byzantine-Robust Distributed Stochastic Learning with Arbitrary Number of Corrupted Workers

被引:1
|
作者
Jian Xu [1 ]
Tong, Xinyi [1 ]
Huang, Shao-Lun [1 ]
机构
[1] Tsinghua Univ, Tsinghua Berkeley Shenzhen Inst, Beijing, Peoples R China
基金
国家重点研发计划;
关键词
Distributed Learning; Communication-Efficient; Byzantine-Robust; Straggler-Resilient;
D O I
10.1109/ICC45855.2022.9838792
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
Distributed implementations of gradient-based algorithms have been essential for training large machine learning models on massive datasets. However, distributed learning algorithms are confronted with several challenges, including communication costs, straggler issues, and attacks from Byzantine adversaries. Existing works on attack-resilient distributed learning, e.g., the coordinate-wise median of gradients, usually neglect communication and/or straggler issues, and fail to defend against well-crafted attacks. Moreover, those methods are ineffective when more than half of workers are corrupted by a Byzantine adversary. To tackle those challenges simultaneously, we develop a robust gradient aggregation framework that is compatible with gradient compression and straggler mitigation techniques. Our proposed framework requires the parameter server to maintain an honest gradient as a reference at each iteration, thus can compute trust-score and similarity for each received gradient and tolerate arbitrary number of corrupted workers. We also provide convergence analysis of our method for non-convex optimization problems. Finally, experiments of image classification task on Fashion-MNIST dataset are conducted under various Byzantine attacks and gradient sparsification operations, and the numerical results demonstrate the effectiveness of our proposed strategy.
引用
收藏
页码:5415 / 5420
页数:6
相关论文
共 50 条
  • [1] Communication-Efficient and Byzantine-Robust Distributed Learning
    Ghosh, Avishek
    Maity, Raj Kumar
    Kadhe, Swanand
    Mazumdar, Arya
    Ramchandran, Kannan
    2020 INFORMATION THEORY AND APPLICATIONS WORKSHOP (ITA), 2020,
  • [2] Communication-Efficient and Byzantine-Robust Distributed Learning with Error Feedback
    Ghosh A.
    Maity R.K.
    Kadhe S.
    Mazumdar A.
    Ramchandran K.
    IEEE Journal on Selected Areas in Information Theory, 2021, 2 (03): : 942 - 953
  • [3] Communication-efficient and Byzantine-robust distributed learning with statistical guarantee
    Zhou, Xingcai
    Chang, Le
    Xu, Pengfei
    Lv, Shaogao
    PATTERN RECOGNITION, 2023, 137
  • [4] Byzantine-Robust and Communication-Efficient Personalized Federated Learning
    Zhang, Jiaojiao
    He, Xuechao
    Huang, Yue
    Ling, Qing
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2025, 73 : 26 - 39
  • [5] Communication-Efficient and Byzantine-Robust Differentially Private Federated Learning
    Li, Min
    Xiao, Di
    Liang, Jia
    Huang, Hui
    IEEE COMMUNICATIONS LETTERS, 2022, 26 (08) : 1725 - 1729
  • [6] STOCHASTIC ADMM FOR BYZANTINE-ROBUST DISTRIBUTED LEARNING
    Lin, Feng
    Ling, Qing
    Li, Weiyu
    Xiong, Zhiwei
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 3172 - 3176
  • [7] Communication-Efficient and Byzantine-Robust Federated Learning for Mobile Edge Computing Networks
    Zhang, Zhuangzhuang
    Wl, Libing
    He, Debiao
    Li, Jianxin
    Cao, Shuqin
    Wu, Xianfeng
    IEEE NETWORK, 2023, 37 (04): : 112 - 119
  • [8] BYZANTINE-ROBUST AND COMMUNICATION-EFFICIENT DISTRIBUTED NON-CONVEX LEARNING OVER NON-IID DATA
    He, Xuechao
    Zhu, Heng
    Ling, Qing
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 5223 - 5227
  • [9] C-RSA: Byzantine-robust and communication-efficient distributed learning in the non-convex and non-IID regime
    He, Xuechao
    Zhu, Heng
    Ling, Qing
    SIGNAL PROCESSING, 2023, 213
  • [10] CB-DSL: Communication-Efficient and Byzantine-Robust Distributed Swarm Learning on Non-i.i.d. Data
    Fan, Xin
    Wang, Yue
    Huo, Yan
    Tian, Zhi
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2024, 10 (01) : 322 - 334