High-Dimensional Stochastic Gradient Quantization for Communication-Efficient Edge Learning

被引:68
|
作者
Du, Yuqing [1 ]
Yang, Sheng [2 ]
Huang, Kaibin [1 ]
机构
[1] Univ Hong Kong, Dept Elect & Elect Engn, Hong Kong, Peoples R China
[2] Univ Paris Saclay, Cent Supelec, Lab Signals & Syst, Gif Sur Yvette 91190, France
关键词
Vector quantization; gradient methods; learning (artificial intelligence); distributed algorithms; LIMITED FEEDBACK; MIMO; MANIFOLDS; DESIGN;
D O I
10.1109/TSP.2020.2983166
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Edge machine learning involves the deployment of learning algorithms at the wireless network edge so as to leverage massive mobile data for enabling intelligent applications. The mainstream edge learning approach, federated learning, has been developed based on distributed gradient descent. Based on the approach, stochastic gradients are computed at edge devices and then transmitted to an edge server for updating a global AI model. Since each stochastic gradient is typically high-dimensional, communication overhead becomes a bottleneck for edge learning. To address this issue, we propose a novel framework of hierarchical gradient quantization and study its effect on the learning performance. First, the framework features a practical hierarchical architecture for decomposing the stochastic gradient into its norm and normalized block gradients, and efficiently quantizes them using a uniform quantizer and a low-dimensional Grassmannian codebook, respectively. Subsequently, the quantized normalized block gradients are scaled and cascaded to yield the quantized normalized stochastic gradient using a socalled hinge vector, which is compressed using another low-dimensional Grassmannian quantizer designed under the criterion of minimum distortion. The other feature of the framework is a bit-allocation scheme for reducing the distortion, which divides the total quantization bits to determine the resolutions of low-dimensional quantizers. The framework is proved to guarantee model convergency by analyzing the convergence rate as a function of quantization bits. Furthermore, by simulation, our design is shown to substantially reduce the communication overhead compared with the state-of-the-art signSGD scheme, while achieving similar learning accuracies.
引用
收藏
页码:2128 / 2142
页数:15
相关论文
共 50 条
  • [1] High-Dimensional Stochastic Gradient Quantization for Communication-Efficient Edge Learning
    Du, Yuqing
    Yang, Sheng
    Huang, Kaibin
    [J]. 2019 7TH IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (IEEE GLOBALSIP), 2019,
  • [2] Communication-efficient Subspace Methods for High-dimensional Federated Learning
    Shi, Zai
    Eryilmaz, Atilla
    [J]. 2021 17TH INTERNATIONAL CONFERENCE ON MOBILITY, SENSING AND NETWORKING (MSN 2021), 2021, : 543 - 550
  • [3] Communication-Efficient Distributed Learning for High-Dimensional Support Vector Machines
    Zhou, Xingcai
    Shen, Hao
    [J]. MATHEMATICS, 2022, 10 (07)
  • [4] AGQFL: Communication-efficient Federated Learning via Automatic Gradient Quantization in Edge Heterogeneous Systems
    Lian, Zirui
    Cao, Jing
    Zuo, Yanru
    Liu, Weihong
    Zhu, Zongwei
    [J]. 2021 IEEE 39TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD 2021), 2021, : 551 - 558
  • [5] Communication-efficient estimation of high-dimensional quantile regression
    Wang, Lei
    Lian, Heng
    [J]. ANALYSIS AND APPLICATIONS, 2020, 18 (06) : 1057 - 1075
  • [6] Communication-Efficient Local Stochastic Gradient Descent for Scalable Deep Learning
    Lee, Sunwoo
    Kang, Qiao
    Agrawal, Ankit
    Choudhary, Alok
    Liao, Wei-keng
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2020, : 718 - 727
  • [7] Communication-Efficient Federated Learning with Adaptive Quantization
    Mao, Yuzhu
    Zhao, Zihao
    Yan, Guangfeng
    Liu, Yang
    Lan, Tian
    Song, Linqi
    Ding, Wenbo
    [J]. ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2022, 13 (04)
  • [8] Communication-Efficient Distributed Estimation of Causal Effects With High-Dimensional Data
    Wang, Xiaohan
    Tong, Jiayi
    Peng, Sida
    Chen, Yong
    Ning, Yang
    [J]. STAT, 2024, 13 (03):
  • [9] Distributed Hierarchical Temporal Graph Learning for Communication-Efficient High-Dimensional Industrial IoT Modeling
    Li, Fangyu
    Lin, Junnuo
    Wang, Yu
    Du, Yongping
    Han, Honggui
    [J]. IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (17): : 28578 - 28590
  • [10] QSGD: Communication-Efficient SGD via Gradient Quantization and Encoding
    Alistarh, Dan
    Grubic, Demjan
    Li, Jerry Z.
    Tomioka, Ryota
    Vojnovic, Milan
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30