Communication-Efficient and Byzantine-Robust Distributed Stochastic Learning with Arbitrary Number of Corrupted Workers

被引:1
|
作者
Jian Xu [1 ]
Tong, Xinyi [1 ]
Huang, Shao-Lun [1 ]
机构
[1] Tsinghua Univ, Tsinghua Berkeley Shenzhen Inst, Beijing, Peoples R China
基金
国家重点研发计划;
关键词
Distributed Learning; Communication-Efficient; Byzantine-Robust; Straggler-Resilient;
D O I
10.1109/ICC45855.2022.9838792
中图分类号
TN [电子技术、通信技术];
学科分类号
0809 ;
摘要
Distributed implementations of gradient-based algorithms have been essential for training large machine learning models on massive datasets. However, distributed learning algorithms are confronted with several challenges, including communication costs, straggler issues, and attacks from Byzantine adversaries. Existing works on attack-resilient distributed learning, e.g., the coordinate-wise median of gradients, usually neglect communication and/or straggler issues, and fail to defend against well-crafted attacks. Moreover, those methods are ineffective when more than half of workers are corrupted by a Byzantine adversary. To tackle those challenges simultaneously, we develop a robust gradient aggregation framework that is compatible with gradient compression and straggler mitigation techniques. Our proposed framework requires the parameter server to maintain an honest gradient as a reference at each iteration, thus can compute trust-score and similarity for each received gradient and tolerate arbitrary number of corrupted workers. We also provide convergence analysis of our method for non-convex optimization problems. Finally, experiments of image classification task on Fashion-MNIST dataset are conducted under various Byzantine attacks and gradient sparsification operations, and the numerical results demonstrate the effectiveness of our proposed strategy.
引用
收藏
页码:5415 / 5420
页数:6
相关论文
共 50 条
  • [31] Toward Byzantine-Robust Distributed Learning for Sentiment Classification on Social Media Platform
    Zhang, Heyi
    Wu, Jun
    Pan, Qianqian
    Bashir, Ali Kashif
    Omar, Marwan
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024, : 1 - 11
  • [32] Robust communication-efficient decentralized learning with heterogeneity
    Zhang, Xiao
    Wang, Yangyang
    Chen, Shuzhen
    Wang, Cui
    Yu, Dongxiao
    Cheng, Xiuzhen
    JOURNAL OF SYSTEMS ARCHITECTURE, 2023, 141
  • [33] More communication-efficient distributed sparse learning
    Zhou, Xingcai
    Yang, Guang
    INFORMATION SCIENCES, 2024, 668
  • [34] BYZANTINE-ROBUST STOCHASTIC GRADIENT DESCENT FOR DISTRIBUTED LOW-RANK MATRIX COMPLETION
    He, Xuechao
    Ling, Qing
    Chen, Tianyi
    2019 IEEE DATA SCIENCE WORKSHOP (DSW), 2019, : 322 - 326
  • [35] Communication-efficient distributed eigenspace estimation with arbitrary node failures
    Charisopoulos, Vasileios
    Damle, Anil
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [36] More communication-efficient distributed sparse learning
    Zhou, Xingcai
    Yang, Guang
    Information Sciences, 2024, 668
  • [37] TrustDDL: A Privacy-Preserving Byzantine-Robust Distributed Deep Learning Framework
    Nikiel, Rene Klaus
    Mirabi, Meghdad
    Binnig, Carsten
    2024 54TH ANNUAL IEEE/IFIP INTERNATIONAL CONFERENCE ON DEPENDABLE SYSTEMS AND NETWORKS WORKSHOPS, DSN-W 2024, 2024, : 55 - 62
  • [38] Byzantine-Robust Distributed Online Learning: Taming Adversarial Participants in An Adversarial Environment
    Dong, Xingrong
    Wu, Zhaoxian
    Ling, Qing
    Tian, Zhi
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2024, 72 : 235 - 248
  • [39] Byzantine-robust decentralized stochastic optimization with stochastic gradient noise-independent learning error
    Peng, Jie
    Li, Weiyu
    Ling, Qing
    SIGNAL PROCESSING, 2024, 219
  • [40] Efficient Byzantine-Robust and Privacy-Preserving Federated Learning on Compressive Domain
    Hu, Guiqiang
    Li, Hongwei
    Fan, Wenshu
    Zhang, Yushu
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (04): : 7116 - 7127