FedQClip: Accelerating Federated Learning via Quantized Clipped SGD

被引:1
|
作者
Qu, Zhihao [1 ]
Jia, Ninghui [1 ]
Ye, Baoliu [2 ]
Hu, Shihong [1 ]
Guo, Song [3 ]
机构
[1] Hohai Univ, Key Lab Water Big Data Technol, Minist Water Resources, Nanjing 211100, Peoples R China
[2] Nanjing Univ, Dept Comp Sci & Technol, State Key Lab Novel Software Technol, Nanjing 210023, Peoples R China
[3] Hong Kong Univ Sci & Technol, Dept Comp Sci & Engn, Kowloon, Hong Kong 999077, Peoples R China
基金
中国国家自然科学基金;
关键词
Convergence; Quantization (signal); Training; Servers; Computers; Computational modeling; Optimization methods; Distance learning; Costs; Computer aided instruction; Federated learning; communication compression; clipped SGD; optimization analysis;
D O I
10.1109/TC.2024.3477972
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Federated Learning (FL) has emerged as a promising technique for collaboratively training machine learning models among multiple participants while preserving privacy-sensitive data. However, the conventional parameter server architecture presents challenges in terms of communication overhead when employing iterative optimization methods such as Stochastic Gradient Descent (SGD). Although communication compression techniques can reduce the traffic cost of FL during each training round, they often lead to degraded convergence rates, mainly due to compression errors and data heterogeneity. To address these issues, this paper presents FedQClip, an innovative approach that combines quantization and Clipped SGD. FedQClip leverages an adaptive step size inversely proportional to the l(2) norm of the gradient, effectively mitigating the negative impacts of quantized errors. Additionally, clipped operations can be applied locally and globally to further expedite training. Theoretical analyses provide evidence that, even under the settings of Non-IID (non-independent and identically distributed) data, FedQClip achieves a convergence rate of O(1/root T), effectively addressing the convergence degradation caused by compression errors. Furthermore, our theoretical analysis highlights the importance of selecting an appropriate number of local updates to enhance the convergence of FL training. Through extensive experiments, we demonstrate that FedQClip outperforms state-of-the-art methods in terms of communication efficiency and convergence rate.
引用
收藏
页码:717 / 730
页数:14
相关论文
共 50 条
  • [21] Generalizable Reconstruction for Accelerating MR Imaging via Federated Learning With Neural Architecture Search
    Wu, Ruoyou
    Li, Cheng
    Zou, Juan
    Liu, Xinfeng
    Zheng, Hairong
    Wang, Shanshan
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2025, 44 (01) : 106 - 117
  • [22] Accelerating Federated Learning With a Global Biased Optimiser
    Mills, Jed
    Hu, Jia
    Min, Geyong
    Jin, Rui
    Zheng, Siwei
    Wang, Jin
    IEEE TRANSACTIONS ON COMPUTERS, 2023, 72 (06) : 1804 - 1814
  • [23] Accelerating Federated Learning at Programmable User Plane Function via In-Network Aggregation
    Bae, Chanbin
    Lee, Hochan
    Pack, Sangheon
    Ji, Youngmin
    38TH INTERNATIONAL CONFERENCE ON INFORMATION NETWORKING, ICOIN 2024, 2024, : 218 - 220
  • [24] Quantized Distributed Federated Learning for Industrial Internet of Things
    Ma, Teng
    Wang, Haibo
    Li, Chong
    IEEE INTERNET OF THINGS JOURNAL, 2023, 10 (04) : 3027 - 3036
  • [25] Distributed Quantized Transmission and Fusion for Federated Machine Learning
    Kandelusy, Omid Moghimi
    Brinton, Christopher G.
    Kim, Taejoon
    2023 IEEE 98TH VEHICULAR TECHNOLOGY CONFERENCE, VTC2023-FALL, 2023,
  • [26] Anchor Model-Based Hybrid Hierarchical Federated Learning With Overlap SGD
    Manjang, Ousman
    Zhai, Yanlong
    Shen, Jun
    Tchaye-Kondi, Jude
    Zhu, Liehuang
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2024, 23 (12) : 12540 - 12557
  • [27] Bias-Variance Reduced Local SGD for Less Heterogeneous Federated Learning
    Murata, Tomoya
    Suzuki, Taiji
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [28] Workie-Talkie: Accelerating Federated Learning by Overlapping Computing and Communications via Contrastive Regularization
    Chen, Rui
    Wan, Qiyu
    Prakash, Pavana
    Zhang, Lan
    Yuan, Xu
    Gong, Yanmin
    Fu, Xin
    Pan, Miao
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 16953 - 16963
  • [29] Accelerating Wireless Federated Learning via Nesterov's Momentum and Distributed Principal Component Analysis
    Dong, Yanjie
    Wang, Luya
    Wang, Jia
    Hu, Xiping
    Zhang, Haijun
    Yu, Fei Richard
    Leung, Victor C. M.
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2024, 23 (06) : 5938 - 5952
  • [30] Early Detection of Diabetes Mellitus Using Differentially Private SGD in Federated Learning
    Dolo, Bakary
    Loukil, Faiza
    Boukadi, Khouloud
    2022 IEEE/ACS 19TH INTERNATIONAL CONFERENCE ON COMPUTER SYSTEMS AND APPLICATIONS (AICCSA), 2022,