How Asynchronous can Federated Learning Be?

被引:10
|
作者
Su, Ningxin [1 ]
Li, Baochun [1 ]
机构
[1] Univ Toronto, Dept Elect & Comp Engn, Toronto, ON, Canada
关键词
D O I
10.1109/IWQoS54832.2022.9812885
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
As a practical paradigm designed to involve large numbers of edge devices in distributed training of deep learning models, federated learning has witnessed a significant amount of research attention in the recent years. Yet, most existing mechanisms on federated learning assumed either fully synchronous or asynchronous communication strategies between clients and the federated learning server. Existing designs that were partially asynchronous in their communication were simple heuristics, and were evaluated using the number of communication rounds or updates required for convergence, rather than the wall-clock time in practice. In this paper, we seek to explore the entire design space between fully synchronous and asynchronous mechanisms of communication. Based on insights from our exploration, we propose PORT, a new partially asynchronous mechanism designed to allow fast clients to aggregate asynchronously, yet without waiting excessively for the slower ones. In addition, PORT is designed to adjust the aggregation weights based on both the staleness and divergence of model updates, with provable convergence guarantees. We have implemented PORT and its leading competitors in PLATO, an open-source scalable federated learning research framework designed from the ground up to emulate real-world scenarios. With respect to the wall-clock time it takes for converging to the target accuracy, PORT outperformed its closest competitor, FedBuff, by up to 40% in our experiments.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Adaptive asynchronous federated learning
    Lu, Renhao
    Zhang, Weizhe
    Li, Qiong
    He, Hui
    Zhong, Xiaoxiong
    Yang, Hongwei
    Wang, Desheng
    Xu, Zenglin
    Alazab, Mamoun
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2024, 152 : 193 - 206
  • [2] Asynchronous Federated Learning for Tactical Autonomy
    Tankard, Earl
    Hagos, Desta Haileselassie
    Rawat, Danda B.
    ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING FOR MULTI-DOMAIN OPERATIONS APPLICATIONS V, 2023, 12538
  • [3] Federated Learning with Buffered Asynchronous Aggregation
    Nguyen, John
    Malik, Kshitiz
    Zhan, Hongyuan
    Yousefpour, Ashkan
    Rabbat, Michael
    Malek, Mani
    Huba, Dzmitry
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [4] Asynchronous federated learning on heterogeneous devices: A survey
    Xu, Chenhao
    Qu, Youyang
    Xiang, Yong
    Gao, Longxiang
    COMPUTER SCIENCE REVIEW, 2023, 50
  • [5] Asynchronous Decentralized Federated Learning for Heterogeneous Devices
    Liao, Yunming
    Xu, Yang
    Xu, Hongli
    Chen, Min
    Wang, Lun
    Qiao, Chunming
    IEEE-ACM TRANSACTIONS ON NETWORKING, 2024, 32 (05) : 4535 - 4550
  • [6] Hierarchical Optimization for Asynchronous Vertical Federated Learning
    Li, Xinchao
    Zhang, Zhixian
    Yang, Shiyou
    Zhou, Xuhua
    HUMAN-CENTRIC COMPUTING AND INFORMATION SCIENCES, 2025, 15
  • [7] When Blockchain Meets Asynchronous Federated Learning
    Jing, Rui
    Chen, Wei
    Wu, Xiaoxin
    Wang, Zehua
    Tian, Zijian
    Zhang, Fan
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT IX, ICIC 2024, 2024, 14870 : 199 - 207
  • [8] Efficient Asynchronous Federated Learning for AUV Swarm
    Meng, Zezhao
    Li, Zhi
    Hou, Xiangwang
    Du, Jun
    Chen, Jianrui
    Wei, Wei
    SENSORS, 2022, 22 (22)
  • [9] Efficient asynchronous federated learning with sparsification and quantization
    Jia, Juncheng
    Liu, Ji
    Zhou, Chendi
    Tian, Hao
    Dong, Mianxiong
    Dou, Dejing
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2024, 36 (09):
  • [10] Wireless Federated Learning With Asynchronous and Quantized Updates
    Huang, Peishan
    Li, Dong
    Yan, Zhigang
    IEEE COMMUNICATIONS LETTERS, 2023, 27 (09) : 2393 - 2397