Communication Efficient and Byzantine Tolerant Distributed Learning

被引:0
|
作者
Ghosh, Avishek [1 ]
Maity, Raj Kumar [2 ]
Kadhe, Swanand [1 ]
Mazumdar, Arya [2 ]
Ramachandran, Kannan [1 ]
机构
[1] Univ Calif Berkeley, Dept Elect Engn & Comp Sci, Berkeley, CA 94720 USA
[2] UMASS Amherst, Coll Informat & Comp Sci, Amherst, MA USA
关键词
D O I
10.1109/isit44484.2020.9174391
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
We develop a communication-efficient distributed learning algorithm that is robust against Byzantine worker machines. We propose and analyze a distributed gradient-descent algorithm that performs a simple thresholding based on gradient norms to mitigate Byzantine failures. We show the (statistical) error-rate of our algorithm matches that of Yin et al., 2018, which uses more complicated schemes (like coordinate-wise median or trimmed mean). Furthermore, for communication efficiency, we consider a generic class of delta-approximate compressors from Karimireddy et al., 2019, that encompasses sign-based compressors and top-k sparsification. Our algorithm uses compressed gradients and gradient norms for aggregation and Byzantine removal respectively. We establish the statistical error rate of the algorithm for arbitrary (convex or non-convex) smooth loss function. We show that, in certain regime of delta, the rate of convergence is not affected by the compression operation. We have experimentally validated our results and shown good performance in convergence for convex (least-square regression) and non-convex (neural network training) problems.
引用
收藏
页码:2545 / 2550
页数:6
相关论文
共 50 条
  • [41] Communication-Efficient Quantum Algorithm for Distributed Machine Learning
    Tang, Hao
    Li, Boning
    Wang, Guoqing
    Xu, Haowei
    Li, Changhao
    Barr, Ariel
    Cappellaro, Paola
    Li, Ju
    PHYSICAL REVIEW LETTERS, 2023, 130 (15)
  • [42] Communication-efficient Distributed Learning for Large Batch Optimization
    Liu, Rui
    Mozafari, Barzan
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [43] Communication-Efficient and Resilient Distributed Q-Learning
    Xie, Yijing
    Mou, Shaoshuai
    Sundaram, Shreyas
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (03) : 3351 - 3364
  • [44] Communication-Efficient Distributed Learning of Discrete Probability Distributions
    Diakonikolas, Ilias
    Grigorescu, Elena
    Li, Jerry
    Natarajan, Abhiram
    Onak, Krzysztof
    Schmidt, Ludwig
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [45] OPTIMAL COMMUNICATION IN NETWORKS WITH RANDOMLY DISTRIBUTED BYZANTINE FAULTS
    BLOUGH, DM
    PELC, A
    NETWORKS, 1993, 23 (08) : 691 - 701
  • [46] CB-DSL: Communication-Efficient and Byzantine-Robust Distributed Swarm Learning on Non-i.i.d. Data
    Fan, Xin
    Wang, Yue
    Huo, Yan
    Tian, Zhi
    IEEE TRANSACTIONS ON COGNITIVE COMMUNICATIONS AND NETWORKING, 2024, 10 (01) : 322 - 334
  • [47] C-RSA: Byzantine-robust and communication-efficient distributed learning in the non-convex and non-IID regime
    He, Xuechao
    Zhu, Heng
    Ling, Qing
    SIGNAL PROCESSING, 2023, 213
  • [48] Byzantine-Robust Distributed Learning With Compression
    Zhu, Heng
    Ling, Qing
    IEEE TRANSACTIONS ON SIGNAL AND INFORMATION PROCESSING OVER NETWORKS, 2023, 9 : 280 - 294
  • [49] Byzantine fault tolerant execution of long-running distributed applications
    Pallemulle, Sajeeva L.
    Wehrman, Ian
    Goldman, Kenneth J.
    PROCEEDINGS OF THE 18TH IASTED INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED COMPUTING AND SYSTEMS, 2006, : 528 - +
  • [50] Byzantine Resilient Distributed Learning in Multirobot Systems
    Li, Jiani
    Abbas, Waseem
    Shabbir, Mudassir
    Koutsoukos, Xenofon
    IEEE TRANSACTIONS ON ROBOTICS, 2022, 38 (06) : 3550 - 3563