MAC Aware Quantization for Distributed Gradient Descent

被引:6
|
作者
Chang, Wei-Ting [1 ]
Tandon, Ravi [1 ]
机构
[1] Univ Arizona, Dept Elect & Comp Engn, Tucson, AZ 85721 USA
关键词
D O I
10.1109/GLOBECOM42002.2020.9322254
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this work, we study the problem of federated learning (FL), where distributed users aim to jointly train a machine learning model with the help of a parameter server (PS). In each iteration of FL, users compute local gradients, followed by transmission of the quantized gradients for subsequent aggregation and model updates at PS. One of the challenges of FL is that of communication overhead due to FL's iterative nature and large model sizes. One recent direction to alleviate communication bottleneck in FL is to let users communicate simultaneously over a multiple access channel (MAC), possibly making better use of the communication resources. In this paper, we consider the problem of FL over a MAC. We focus on the design of digital gradient transmission schemes over a MAC, where gradients at each user are first quantized, and then transmitted over a MAC to be decoded individually at the PS. When designing digital FL schemes over MACs, there are new opportunities to assign different amount of resources (e.g., rate or bandwidth) to different users based on a) the informativeness of the gradients at users, and b) the underlying channel conditions. We propose a stochastic gradient quantization scheme, where the quantization parameters are optimized based on the capacity region of the MAC. We show that such channel aware quantization for FL outperforms uniform quantization, particularly when users experience different channel conditions, and when have gradients with varying levels of informativeness.
引用
收藏
页数:6
相关论文
共 50 条
  • [31] A Mirror Descent-Based Algorithm for Corruption-Tolerant Distributed Gradient Descent
    Wang, Shuche
    Tan, Vincent Y. F.
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2025, 73 : 827 - 842
  • [32] GRADIENT ALGORITHM FOR QUANTIZATION LEVELS IN DISTRIBUTED DETECTION SYSTEMS
    HELSTROM, CW
    IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS, 1995, 31 (01) : 390 - 398
  • [33] ACCELERATING DISTRIBUTED DEEP LEARNING BY ADAPTIVE GRADIENT QUANTIZATION
    Guo, Jinrong
    Liu, Wantao
    Wang, Wang
    Han, Jizhong
    Li, Ruixuan
    Lu, Yijun
    Hu, Songlin
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 1603 - 1607
  • [34] Quantized Gradient-Descent Algorithm for Distributed Resource Allocation
    Zhou, Hongbing
    Yu, Weiyong
    Yi, Peng
    Hong, Yiguang
    UNMANNED SYSTEMS, 2019, 7 (02) : 119 - 136
  • [35] Introducing distributed gradient descent distance vector routing (DGDDVR)
    Herbert, S. J.
    ELECTRONICS LETTERS, 2016, 52 (23) : 1963 - 1965
  • [36] ROBUST DISTRIBUTED GRADIENT DESCENT WITH ARBITRARY NUMBER OF BYZANTINE ATTACKERS
    Cao, Xinyang
    Lai, Lifeng
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 6373 - 6377
  • [37] A Sharp Estimate on the Transient Time of Distributed Stochastic Gradient Descent
    Pu, Shi
    Olshevsky, Alex
    Paschalidis, Ioannis Ch
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2022, 67 (11) : 5900 - 5915
  • [38] A Distributed Optimal Control Problem with Averaged Stochastic Gradient Descent
    Sun, Qi
    Du, Qiang
    COMMUNICATIONS IN COMPUTATIONAL PHYSICS, 2020, 27 (03) : 753 - 774
  • [39] Scaling Stratified Stochastic Gradient Descent for Distributed Matrix Completion
    Abubaker N.
    Karsavuran M.O.
    Aykanat C.
    IEEE Transactions on Knowledge and Data Engineering, 2023, 35 (10) : 10603 - 10615
  • [40] Securing Distributed Gradient Descent in High Dimensional Statistical Learning
    Su, Lili
    Xu, Jiaming
    PROCEEDINGS OF THE ACM ON MEASUREMENT AND ANALYSIS OF COMPUTING SYSTEMS, 2019, 3 (01)