BASGD: Buffered Asynchronous SGD for Byzantine Learning

被引:0
|
作者
Yang, Yi-Rui [1 ]
Li, Wu-Jun [1 ]
机构
[1] Nanjing Univ, Dept Comp Sci & Technol, Natl Key Lab Novel Software Technol, Nanjing, Peoples R China
基金
国家重点研发计划;
关键词
PARALLEL;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Distributed learning has become a hot research topic due to its wide application in cluster-based large-scale learning, federated learning, edge computing and so on. Most traditional distributed learning methods typically assume no failure or attack. However, many unexpected cases, such as communication failure and even malicious attack, may happen in real applications. Hence, Byzantine learning (BL), which refers to distributed learning with failure or attack, has recently attracted much attention. Most existing BL methods are synchronous, which are impractical in some applications due to heterogeneous or offline workers. In these cases, asynchronous BL (ABL) is usually preferred. In this paper, we propose a novel method, called buffered asynchronous stochastic gradient descent (BASGD), for ABL. To the best of our knowledge, BASGD is the first ABL method that can resist malicious attack without storing any instances on server. Compared with those methods which need to store instances on server, BASGD has a wider scope of application. BASGD is proved to be convergent, and be able to resist failure or attack. Empirical results show that BASGD significantly outperforms vanilla asynchronous stochastic gradient descent (ASGD) and other ABL baselines when there exists failure or attack on workers.
引用
下载
收藏
页数:11
相关论文
共 50 条
  • [31] Asynchronous Byzantine Agreement with optimal resilience
    Patra, Arpita
    Choudhury, Ashish
    Rangan, C. Pandu
    DISTRIBUTED COMPUTING, 2014, 27 (02) : 111 - 146
  • [32] On optimal probabilistic asynchronous Byzantine agreement
    Shareef, Amjed
    Rangan, C. Pandu
    DISTRIBUTED COMPUTING AND NETWORKING, PROCEEDINGS, 2008, 4904 : 86 - 98
  • [33] Asynchronous Byzantine Agreement with Subquadratic Communication
    Blum, Erica
    Katz, Jonathan
    Liu-Zhang, Chen-Da
    Loss, Julian
    THEORY OF CRYPTOGRAPHY, TCC 2020, PT I, 2020, 12550 : 353 - 380
  • [34] Detecting Malicious Gradients from Asynchronous SGD on Variational Autoencoder
    Gu, Zhipin
    Yang, Yuexiang
    Shi, Heyuan
    2021 40TH INTERNATIONAL SYMPOSIUM ON RELIABLE DISTRIBUTED SYSTEMS (SRDS 2021), 2021, : 321 - 330
  • [35] Learning with SGD and Random Features
    Carratino, Luigi
    Rudi, Alessandro
    Rosasco, Lorenzo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [36] IS-ASGD: Accelerating Asynchronous SGD using Importance Sampling
    Wang, Fei
    Gao, Xiaofeng
    Ye, Jun
    Chen, Guihai
    PROCEEDINGS OF THE 47TH INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING, 2018,
  • [37] Stability-Based Generalization Analysis of the Asynchronous Decentralized SGD
    Deng, Xiaoge
    Sun, Tao
    Li, Shengwei
    Li, Dongsheng
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 6, 2023, : 7340 - 7348
  • [38] Predicting Statistics of Asynchronous SGD Parameters for a Large-Scale Distributed Deep Learning System on GPU Supercomputers
    Oyama, Yosuke
    Nomura, Akihiro
    Sato, Ikuro
    Nishimura, Hiroki
    Tamatsu, Yukimasa
    Matsuoka, Satoshi
    2016 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2016, : 66 - 75
  • [39] AsGrad: A Sharp Unified Analysis of Asynchronous-SGD Algorithms
    Islamov, Rustem
    Safaryan, Mher
    Alistarh, Dan
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [40] On the convergence analysis of asynchronous SGD for solving consistent linear systems
    Sahu, Atal Narayan
    Dutta, Aritra
    Tiwari, Aashutosh
    Richtarik, Peter
    LINEAR ALGEBRA AND ITS APPLICATIONS, 2023, 663 : 1 - 31