A Communication-Efficient Hierarchical Federated Learning Framework via Shaping Data Distribution at Edge

被引:7
|
作者
Deng, Yongheng [1 ]
Lyu, Feng [2 ]
Xia, Tengxi [1 ]
Zhou, Yuezhi [3 ]
Zhang, Yaoxue [1 ,3 ]
Ren, Ju [1 ,3 ]
Yang, Yuanyuan [4 ]
机构
[1] Tsinghua Univ, Beijing Natl Res Ctr Informat Sci & Technol BNRist, Dept Comp Sci & Technol, Beijing 100084, Peoples R China
[2] Cent South Univ, Sch Comp Sci & Engn, Changsha 410083, Peoples R China
[3] Zhongguancun Lab, Beijing 100084, Peoples R China
[4] SUNY Stony Brook, Dept Elect & Comp Engn, Stony Brook, NY 11794 USA
关键词
Costs; Data models; Servers; Computational modeling; Training data; Federated learning; Distributed databases; Hierarchical federated learning; communication efficiency; edge computing; distributed edge intelligence; RESOURCE-ALLOCATION;
D O I
10.1109/TNET.2024.3363916
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning (FL) enables collaborative model training over distributed computing nodes without sharing their privacy-sensitive raw data. However, in FL, iterative exchanges of model updates between distributed nodes and the cloud server can result in significant communication cost, especially when the data distributions at distributed nodes are imbalanced with requiring more rounds of iterations. In this paper, with our in-depth empirical studies, we disclose that extensive cloud aggregations can be avoided without compromising the learning accuracy if frequent aggregations can be enabled at edge network. To this end, we shed light on the hierarchical federated learning (HFL) framework, where a subset of distributed nodes can play as edge aggregators to support edge aggregations. Under the HFL framework, we formulate a communication cost minimization (CCM) problem to minimize the total communication cost required for model learning with a target accuracy by making decisions on edge aggragator selection and node-edge associations. Inspired by our data-driven insights that the potential of HFL lies in the data distribution at edge aggregators, we propose ShapeFL, i.e., SHaping dAta distRibution at Edge, to transform and solve the CCM problem. In ShapeFL, we divide the original problem into two sub-problems to minimize the per-round communication cost and maximize the data distribution diversity of edge aggregator data, respectively, and devise two light-weight algorithms to solve them accordingly. Extensive experiments are carried out based on several opened datasets and real-world network topologies, and the results demonstrate the efficacy of ShapeFL in terms of both learning accuracy and communication efficiency.
引用
收藏
页码:2600 / 2615
页数:16
相关论文
共 50 条
  • [31] Communication-efficient clustered federated learning via model distance
    Mao Zhang
    Tie Zhang
    Yifei Cheng
    Changcun Bao
    Haoyu Cao
    Deqiang Jiang
    Linli Xu
    Machine Learning, 2024, 113 : 3869 - 3888
  • [32] FedCO: Communication-Efficient Federated Learning via Clustering Optimization
    Al-Saedi, Ahmed A.
    Boeva, Veselka
    Casalicchio, Emiliano
    FUTURE INTERNET, 2022, 14 (12)
  • [33] Communication-Efficient Adaptive Federated Learning
    Wang, Yujia
    Lin, Lu
    Chen, Jinghui
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [34] FedHD: Communication-efficient federated learning from hybrid data
    Gao, Haiqing
    Ge, Songyang
    Chang, Tsung-Hui
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2023, 360 (12): : 8416 - 8454
  • [35] FLCP: federated learning framework with communication-efficient and privacy-preserving
    Yang, Wei
    Yang, Yuan
    Xi, Yingjie
    Zhang, Hailong
    Xiang, Wei
    APPLIED INTELLIGENCE, 2024, 54 (9-10) : 6816 - 6835
  • [36] Communication-efficient Federated Learning Framework with Parameter-Ordered Dropout
    Li, Qichen
    Shao, Sujie
    Yang, Chao
    Chen, Jiewei
    Qi, Feng
    Guo, Shaoyong
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 1195 - 1200
  • [37] Communication-efficient semi-synchronous hierarchical federated learning with balanced training in heterogeneous IoT edge environments
    Herabad, Mohammadsadeq Garshasbi
    INTERNET OF THINGS, 2023, 21
  • [38] Communication-Efficient Federated Learning via Regularized Sparse Random Networks
    Mestoukirdi, Mohamad
    Esrafilian, Omid
    Gesbert, David
    Li, Qianrui
    Gresset, Nicolas
    IEEE COMMUNICATIONS LETTERS, 2024, 28 (07) : 1574 - 1578
  • [39] Communication-Efficient and Attack-Resistant Federated Edge Learning With Dataset Distillation
    Zhou, Yanlin
    Ma, Xiyao
    Wu, Dapeng
    Li, Xiaolin
    IEEE TRANSACTIONS ON CLOUD COMPUTING, 2023, 11 (03) : 2517 - 2528
  • [40] Communication-Efficient Vertical Federated Learning via Compressed Error Feedback
    Valdeira, Pedro
    Xavier, Joao
    Soares, Claudia
    Chi, Yuejie
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2025, 73 : 1065 - 1080