BASGD: Buffered Asynchronous SGD for Byzantine Learning

被引:0
|
作者
Yang, Yi-Rui [1 ]
Li, Wu-Jun [1 ]
机构
[1] Nanjing Univ, Dept Comp Sci & Technol, Natl Key Lab Novel Software Technol, Nanjing, Peoples R China
基金
国家重点研发计划;
关键词
PARALLEL;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Distributed learning has become a hot research topic due to its wide application in cluster-based large-scale learning, federated learning, edge computing and so on. Most traditional distributed learning methods typically assume no failure or attack. However, many unexpected cases, such as communication failure and even malicious attack, may happen in real applications. Hence, Byzantine learning (BL), which refers to distributed learning with failure or attack, has recently attracted much attention. Most existing BL methods are synchronous, which are impractical in some applications due to heterogeneous or offline workers. In these cases, asynchronous BL (ABL) is usually preferred. In this paper, we propose a novel method, called buffered asynchronous stochastic gradient descent (BASGD), for ABL. To the best of our knowledge, BASGD is the first ABL method that can resist malicious attack without storing any instances on server. Compared with those methods which need to store instances on server, BASGD has a wider scope of application. BASGD is proved to be convergent, and be able to resist failure or attack. Empirical results show that BASGD significantly outperforms vanilla asynchronous stochastic gradient descent (ASGD) and other ABL baselines when there exists failure or attack on workers.
引用
下载
收藏
页数:11
相关论文
共 50 条
  • [1] Buffered Asynchronous SGD for Byzantine Learning
    Yang, Yi-Rui
    Li, Wu-Jun
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [2] Asynchronous Byzantine Machine Learning (the case of SGD)
    Damaskinos, Georgios
    El Mhamdi, El Mandi
    Guerraoui, Rachid
    Patra, Rhicheek
    Taziki, Mahsa
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [3] Federated Learning with Buffered Asynchronous Aggregation
    Nguyen, John
    Malik, Kshitiz
    Zhan, Hongyuan
    Yousefpour, Ashkan
    Rabbat, Michael
    Malek, Mani
    Huba, Dzmitry
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [4] Sharper Convergence Guarantees for Asynchronous SGD for Distributed and Federated Learning
    Koloskova, Anastasia
    Stich, Sebastian U.
    Jaggi, Martin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [5] Unbounded Gradients in Federated Learning with Buffered Asynchronous Aggregation
    Toghani, Mohammad Taha
    Uribe, Cesar A.
    2022 58TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2022,
  • [6] Asynchronous SGD with stale gradient dynamic adjustment for deep learning training
    Tan, Tao
    Xie, Hong
    Xia, Yunni
    Shi, Xiaoyu
    Shang, Mingsheng
    INFORMATION SCIENCES, 2024, 681
  • [7] Asynchronous SGD Beats Minibatch SGD Under Arbitrary Delays
    Mishchenko, Konstantin
    Bach, Francis
    Even, Mathieu
    Woodworth, Blake
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [8] AFLGuard: Byzantine-robust Asynchronous Federated Learning
    Fang, Minghong
    Liu, Jia
    Gong, Neil Zhenqiang
    Bentley, Elizabeth S.
    PROCEEDINGS OF THE 38TH ANNUAL COMPUTER SECURITY APPLICATIONS CONFERENCE, ACSAC 2022, 2022, : 632 - 646
  • [9] Collaborative Learning in the Jungle (Decentralized, Byzantine, Heterogeneous, Asynchronous and Nonconvex Learning)
    El-Mhamdi, El-Mandi
    Farhadkhani, Sadegh
    Guerraoui, Rachid
    Guirguis, Arsany
    Le-Nguyen Hoang
    Rouault, Sebastien
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [10] Throughput Prediction of Asynchronous SGD in TensorFlow
    Li, Zhuojin
    Yan, Wumo
    Paolieri, Marco
    Golubchik, Leana
    PROCEEDINGS OF THE ACM/SPEC INTERNATIONAL CONFERENCE ON PERFORMANCE ENGINEERING (ICPE'20), 2020, : 76 - 87