How Asynchronous can Federated Learning Be?

被引:10
|
作者
Su, Ningxin [1 ]
Li, Baochun [1 ]
机构
[1] Univ Toronto, Dept Elect & Comp Engn, Toronto, ON, Canada
关键词
D O I
10.1109/IWQoS54832.2022.9812885
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
As a practical paradigm designed to involve large numbers of edge devices in distributed training of deep learning models, federated learning has witnessed a significant amount of research attention in the recent years. Yet, most existing mechanisms on federated learning assumed either fully synchronous or asynchronous communication strategies between clients and the federated learning server. Existing designs that were partially asynchronous in their communication were simple heuristics, and were evaluated using the number of communication rounds or updates required for convergence, rather than the wall-clock time in practice. In this paper, we seek to explore the entire design space between fully synchronous and asynchronous mechanisms of communication. Based on insights from our exploration, we propose PORT, a new partially asynchronous mechanism designed to allow fast clients to aggregate asynchronously, yet without waiting excessively for the slower ones. In addition, PORT is designed to adjust the aggregation weights based on both the staleness and divergence of model updates, with provable convergence guarantees. We have implemented PORT and its leading competitors in PLATO, an open-source scalable federated learning research framework designed from the ground up to emulate real-world scenarios. With respect to the wall-clock time it takes for converging to the target accuracy, PORT outperformed its closest competitor, FedBuff, by up to 40% in our experiments.
引用
收藏
页数:11
相关论文
共 50 条
  • [21] On the Decentralization of Blockchain-enabled Asynchronous Federated Learning
    Wilhelmi, Francesc
    Guerra, Elia
    Dini, Paolo
    2023 IEEE 9TH INTERNATIONAL CONFERENCE ON NETWORK SOFTWARIZATION, NETSOFT, 2023, : 408 - 413
  • [22] A decentralized asynchronous federated learning framework for edge devices
    Wang, Bin
    Tian, Zhao
    Ma, Jie
    Zhang, Wenju
    She, Wei
    Liu, Wei
    FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2025, 166
  • [23] Staleness aware semi-asynchronous federated learning
    Yu, Miri
    Choi, Jiheon
    Lee, Jaehyun
    Oh, Sangyoon
    JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING, 2024, 93
  • [24] An Asynchronous Federated Learning Mechanism for Edge Network Computing
    Lu X.
    Liao Y.
    Lio P.
    Pan H.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2020, 57 (12): : 2571 - 2582
  • [25] Asynchronous Federated Learning over Wireless Communication Networks
    Wang, Zhongyu
    Zhang, Zhaoyang
    Wang, Jue
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2021), 2021,
  • [26] Federated learning based on asynchronous and adjusted client training
    Dai, Mingjun
    Zhao, Yinglin
    Yuan, Jialong
    Kianoush, Sanaz
    Savazzi, Stefano
    Li, Bingchun
    PHYSICAL COMMUNICATION, 2023, 61
  • [27] Asynchronous Federated Learning System Based on Permissioned Blockchains
    Wang, Rong
    Tsai, Wei-Tek
    SENSORS, 2022, 22 (04)
  • [28] BLADE: Pushing the Performance Envelope of Asynchronous Federated Learning
    Ying, Chen
    Li, Baochun
    Li, Bo
    2024 IEEE/ACM 32ND INTERNATIONAL SYMPOSIUM ON QUALITY OF SERVICE, IWQOS, 2024,
  • [29] Unbounded Gradients in Federated Learning with Buffered Asynchronous Aggregation
    Toghani, Mohammad Taha
    Uribe, Cesar A.
    2022 58TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2022,
  • [30] Asynchronous Federated Learning Through Online Linear Regressions
    Kashima, Taiga
    Amma, Ayako
    Nakayama, Hideki
    IEEE ACCESS, 2024, 12 : 195131 - 195144