Let Them Drop: Scalable and Efficient Federated Learning Solutions Agnostic to Stragglers

被引:0
|
作者
Taiello, Riccardo [1 ]
Onen, Melek [2 ]
Gritti, Clementine [3 ]
Lorenzi, Marco [4 ]
机构
[1] Univ Cote Azur, EURECOM, Inria, Sophia Antipolis, France
[2] EURECOM, Sophia Antipolis, France
[3] Inria, INSA Lyon, Villeurbanne, France
[4] Univ Cote Azur, Inria, Sophia Antipolis, France
关键词
Secure Aggregation; Synchronous and Asynchronous Federated Learning;
D O I
10.1145/3664476.3664488
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Secure Aggregation (SA) stands as a crucial component in modern Federated Learning (FL) systems, facilitating collaborative training of a global machine learning model while protecting the privacy of individual clients' local datasets. Many existing SA protocols described in the FL literature operate synchronously, leading to notable runtime slowdowns due to the presence of stragglers (i.e. late-arriving clients). To address this challenge, one common approach is to consider stragglers as client failures and use SA solutions that are robust against dropouts. While this approach indeed seems to work, it unfortunately affects the performance of the protocol as its cost strongly depends on the dropout ratio and this ratio has increased significantly when taking stragglers into account. Another approach explored in the literature to address stragglers is to introduce asynchronicity into the FL system. Very few SA solutions exist in this setting and currently suffer from high overhead. In this paper, similar to related work, we propose to handle stragglers as client failures but design SA solutions that do not depend on the dropout ratio so that an unavoidable increase on this metric does not affect the performance of the solution. We first introduce Eagle, a synchronous SA scheme designed not to depend on the client failures but on the online users' inputs only. This approach offers better computation and communication costs compared to existing solutions under realistic settings where the number of stragglers is high. We then propose Owl, the first SA solution that is suitable for the asynchronous setting and once again considers online clients' contributions only. We implement both solutions and show that: (i) in a synchronous FL with realistic dropout rates (taking potential stragglers into account), Eagle outperforms the best SA solution, namely Flamingo, by x4; (ii) In the asynchronous setting, Owl exhibits the best performance compared to the state-of-the-art solution LightSecAgg.
引用
收藏
页数:12
相关论文
共 9 条
  • [1] Toward Scalable Wireless Federated Learning: Challenges and Solutions
    Zhou Y.
    Shi Y.
    Zhou H.
    Wang J.
    Fu L.
    Yang Y.
    IEEE Internet of Things Magazine, 2023, 6 (04): : 10 - 16
  • [2] Scalable Video Streaming Solutions Using Federated Learning
    Darwich, Mahmoud
    Khalil, Kasem
    Bayoumi, Magdy
    2024 INTERNATIONAL CONFERENCE ON SMART APPLICATIONS, COMMUNICATIONS AND NETWORKING, SMARTNETS-2024, 2024,
  • [3] Intrinsic Gradient Compression for Scalable and Efficient Federated Learning
    Melas-Kyriazi, Luke
    Wang, Franklyn
    PROCEEDINGS OF THE FIRST WORKSHOP ON FEDERATED LEARNING FOR NATURAL LANGUAGE PROCESSING (FL4NLP 2022), 2022, : 27 - 41
  • [4] Auxo: Efficient Federated Learning via Scalable Client Clustering
    Liu, Jiachen
    Lai, Fan
    Dai, Yinwei
    Akella, Aditya
    Madhyastha, Harsha V.
    Chowdhury, Mosharaf
    PROCEEDINGS OF THE 2023 ACM SYMPOSIUM ON CLOUD COMPUTING, SOCC 2023, 2023, : 125 - 141
  • [5] Communication-efficient and Scalable Decentralized Federated Edge Learning
    Yapp, Austine Zong Han
    Koh, Hong Soo Nicholas
    Lai, Yan Ting
    Kang, Jiawen
    Li, Xuandi
    Ng, Jer Shyuan
    Jiang, Hongchao
    Lim, Wei Yang Bryan
    Xiong, Zehui
    Niyato, Dusit
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 5032 - 5035
  • [6] FedCon: Scalable and Efficient Federated Learning via Contribution-Based Aggregation
    Gao, Wenyu
    Xu, Gaochao
    Meng, Xianqiu
    ELECTRONICS, 2025, 14 (05):
  • [7] FINISH: Efficient and Scalable NMF-Based Federated Learning for Detecting Malware Activities
    Chang, Yu-Wei
    Chen, Hong-Yen
    Han, Chansu
    Morikawa, Tomohiro
    Takahashi, Takeshi
    Lin, Tsung-Nan
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTING, 2023, 11 (04) : 934 - 949
  • [8] Scalable and Resource-Efficient Second-Order Federated Learning via Over-the-Air Aggregation
    Ghalkha, Abdulmomen
    Ben Issaid, Chaouki
    Bennis, Mehdi
    IEEE WIRELESS COMMUNICATIONS LETTERS, 2025, 14 (03) : 716 - 720
  • [9] Mix2SFL: Two-Way Mixup for Scalable, Accurate, and Communication-Efficient Split Federated Learning
    Oh, Seungeun
    Nam, Hyelin
    Park, Jihong
    Vepakomma, Praneeth
    Raskar, Ramesh
    Bennis, Mehdi
    Kim, Seong-Lyun
    IEEE TRANSACTIONS ON BIG DATA, 2024, 10 (03) : 238 - 248