In-Network Computation for Large-Scale Federated Learning Over Wireless Edge Networks

被引:11
|
作者
Dinh, Thinh Quang [1 ]
Nguyen, Diep N. [1 ]
Hoang, Dinh Thai [1 ]
Pham, Tran Vu [2 ]
Dutkiewicz, Eryk [1 ]
机构
[1] Univ Technol Sydney, Sch Elect & Data Engn, Ultimo, NSW 2007, Australia
[2] Ho Chi Minh City Univ Technol HCMUT, VNU HCM, Ho Chi Minh City 70000, Vietnam
基金
澳大利亚研究理事会;
关键词
Computational modeling; Servers; Routing; Training; Network architecture; Machine learning; Stars; Mobile edge computing; federated learning; in-network computation; large-scale distributed learning;
D O I
10.1109/TMC.2022.3190260
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Most conventional Federated Learning (FL) models are using a star network topology where all users aggregate their local models at a single server (e.g., a cloud server). That causes significant overhead in terms of both communications and computing at the server, delaying the training process, especially for large scale FL systems with straggling nodes. This article proposes a novel edge network architecture that enables decentralizing the model aggregation process at the server, thereby significantly reducing the training delay for the whole FL network. Specifically, we design a highly-effective in-network computation framework (INC) consisting of a user scheduling mechanism, an in-network aggregation process (INA) which is designed for both primal- and primal-dual methods in distributed machine learning problems, and a network routing algorithm with theoretical performance bounds. The in-network aggregation process, which is implemented at edge nodes and cloud node, can adapt two typical methods to allow edge networks to effectively solve the distributed machine learning problems. Under the proposed INA, we then formulate a joint routing and resource optimization problem, aiming to minimize the aggregation latency. The problem turns out to be NP-hard, and thus we propose a polynomial time routing algorithm which can achieve near optimal performance with a theoretical bound. Simulation results showed that the proposed algorithm can achieve more than 99% of the optimal solution and reduce the FL training latency, up to 5.6 times w.r.t other baselines. The proposed INC framework can not only help reduce the FL training latency but also significantly decrease cloud's traffic and computing overhead. By embedding the computing/aggregation tasks at the edge nodes and leveraging the multi-layer edge-network architecture, the INC framework can liberate FL from the star topology to enable large-scale FL.
引用
收藏
页码:5918 / 5932
页数:15
相关论文
共 50 条
  • [31] CFL: Cluster Federated Learning in Large-scale Peer-to-Peer Networks
    Chen, Qian
    Wang, Zilong
    Zhou, Yilin
    Chen, Jiawei
    Xiao, Dan
    Lin, Xiaodong
    arXiv, 2022,
  • [32] Decentralized Federated Learning With Asynchronous Parameter Sharing for Large-Scale IoT Networks
    Xie, Haihui
    Xia, Minghua
    Wu, Peiran
    Wang, Shuai
    Huang, Kaibin
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (21): : 34123 - 34139
  • [33] Deploying Federated Learning in Large-Scale Cellular Networks: Spatial Convergence Analysis
    Lin, Zhenyi
    Li, Xiaoyang
    Lau, Vincent K. N.
    Gong, Yi
    Huang, Kaibin
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2022, 21 (03) : 1542 - 1556
  • [34] FedPCC: Parallelism of Communication and Computation for Federated Learning in Wireless Networks
    Zhang, Hong
    Tian, Hao
    Dong, Mianxiong
    Ota, Kaoru
    Jia, Juncheng
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2022, 6 (06): : 1368 - 1377
  • [35] Adaptive Hierarchical Federated Learning Over Wireless Networks
    Xu, Bo
    Xia, Wenchao
    Wen, Wanli
    Liu, Pei
    Zhao, Haitao
    Zhu, Hongbo
    IEEE TRANSACTIONS ON VEHICULAR TECHNOLOGY, 2022, 71 (02) : 2070 - 2083
  • [36] Asynchronous Federated Learning over Wireless Communication Networks
    Wang, Zhongyu
    Zhang, Zhaoyang
    Wang, Jue
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2021), 2021,
  • [37] An Overview of Enabling Federated Learning over Wireless Networks
    Foukalas, Fotis
    Tziouvaras, Athanasios
    Tsiftsis, Theodoros A.
    2021 IEEE INTERNATIONAL MEDITERRANEAN CONFERENCE ON COMMUNICATIONS AND NETWORKING (IEEE MEDITCOM 2021), 2021, : 271 - 276
  • [38] Performance Optimization of Federated Learning over Wireless Networks
    Chen, Mingzhe
    Yang, Zhaohui
    Saad, Walid
    Yin, Changchuan
    Poor, H. Vincent
    Cui, Shuguang
    2019 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2019,
  • [39] Federated Learning Over Energy Harvesting Wireless Networks
    Hamdi, Rami
    Chen, Mingzhe
    Ben Said, Ahmed
    Qaraqe, Marwa
    Poor, H. Vincent
    IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (01) : 92 - 103
  • [40] Serial In-network Processing for Large Stationary Wireless Sensor Networks
    Merzoug, Mohammed Amine
    Boukerche, Azzedine
    Mostefaoui, Ahmed
    PROCEEDINGS OF THE 20TH ACM INTERNATIONAL CONFERENCE ON MODELLING, ANALYSIS AND SIMULATION OF WIRELESS AND MOBILE SYSTEMS (MSWIM'17), 2017, : 153 - 160