In-Network Computation for Large-Scale Federated Learning Over Wireless Edge Networks

被引:11
|
作者
Dinh, Thinh Quang [1 ]
Nguyen, Diep N. [1 ]
Hoang, Dinh Thai [1 ]
Pham, Tran Vu [2 ]
Dutkiewicz, Eryk [1 ]
机构
[1] Univ Technol Sydney, Sch Elect & Data Engn, Ultimo, NSW 2007, Australia
[2] Ho Chi Minh City Univ Technol HCMUT, VNU HCM, Ho Chi Minh City 70000, Vietnam
基金
澳大利亚研究理事会;
关键词
Computational modeling; Servers; Routing; Training; Network architecture; Machine learning; Stars; Mobile edge computing; federated learning; in-network computation; large-scale distributed learning;
D O I
10.1109/TMC.2022.3190260
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Most conventional Federated Learning (FL) models are using a star network topology where all users aggregate their local models at a single server (e.g., a cloud server). That causes significant overhead in terms of both communications and computing at the server, delaying the training process, especially for large scale FL systems with straggling nodes. This article proposes a novel edge network architecture that enables decentralizing the model aggregation process at the server, thereby significantly reducing the training delay for the whole FL network. Specifically, we design a highly-effective in-network computation framework (INC) consisting of a user scheduling mechanism, an in-network aggregation process (INA) which is designed for both primal- and primal-dual methods in distributed machine learning problems, and a network routing algorithm with theoretical performance bounds. The in-network aggregation process, which is implemented at edge nodes and cloud node, can adapt two typical methods to allow edge networks to effectively solve the distributed machine learning problems. Under the proposed INA, we then formulate a joint routing and resource optimization problem, aiming to minimize the aggregation latency. The problem turns out to be NP-hard, and thus we propose a polynomial time routing algorithm which can achieve near optimal performance with a theoretical bound. Simulation results showed that the proposed algorithm can achieve more than 99% of the optimal solution and reduce the FL training latency, up to 5.6 times w.r.t other baselines. The proposed INC framework can not only help reduce the FL training latency but also significantly decrease cloud's traffic and computing overhead. By embedding the computing/aggregation tasks at the edge nodes and leveraging the multi-layer edge-network architecture, the INC framework can liberate FL from the star topology to enable large-scale FL.
引用
收藏
页码:5918 / 5932
页数:15
相关论文
共 50 条
  • [21] Federated Learning-Based In-Network Traffic Analysis on IoT Edge
    Zang, Mingyuan
    Zheng, Changgang
    Koziak, Tomasz
    Zilberman, Noa
    Dittmann, Lars
    2023 IFIP NETWORKING CONFERENCE, IFIP NETWORKING, 2023,
  • [22] On the Connectivity Analysis over Large-Scale Hybrid Wireless Networks
    Yi, Chi
    Wang, Wenye
    2010 PROCEEDINGS IEEE INFOCOM, 2010,
  • [23] Enhanced In-Network Caching for Deep Learning in Edge Networks
    Zhang, Jiaqi
    Liu, Wenjing
    Zhang, Li
    Tian, Jie
    ELECTRONICS, 2024, 13 (23):
  • [24] Client Selection Approach in Support of Clustered Federated Learning over Wireless Edge Networks
    Albaseer, Abdullatif
    Abdallah, Mohamed
    Al-Fuqaha, Ala
    Erbad, Aiman
    2021 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2021,
  • [25] Coded Computing for Low-Latency Federated Learning Over Wireless Edge Networks
    Prakash, Saurav
    Dhakal, Sagar
    Akdeniz, Mustafa Riza
    Yona, Yair
    Talwar, Shilpa
    Avestimehr, Salman
    Himayat, Nageen
    IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2021, 39 (01) : 233 - 250
  • [26] Federated Edge Learning With Misaligned Over-the-Air Computation
    Shao, Yulin
    Gunduz, Deniz
    Liew, Soung Chang
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2022, 21 (06) : 3951 - 3964
  • [27] Federated Edge Learning with Misaligned Over-The-Air Computation
    Shao, Yulin
    Gunduz, Deniz
    Liew, Soung Chang
    SPAWC 2021: 2021 IEEE 22ND INTERNATIONAL WORKSHOP ON SIGNAL PROCESSING ADVANCES IN WIRELESS COMMUNICATIONS (IEEE SPAWC 2021), 2021, : 236 - 240
  • [28] Function Placement for In-network Federated Learning
    Yellas, Nour-El-Houda
    Addis, Bernardetta
    Boumerdassi, Selma
    Riggio, Roberto
    Secci, Stefano
    COMPUTER NETWORKS, 2025, 256
  • [29] In-network Computation for IoT Data Processing with ActiveNDN in Wireless Sensor Networks
    Mekbungwan, Preechai
    Pau, Giovanni
    Kanchanasut, Kanchana
    2022 5TH CONFERENCE ON CLOUD AND INTERNET OF THINGS, CIOT, 2022, : 197 - 204
  • [30] CFL: Cluster Federated Learning in Large-Scale Peer-to-Peer Networks
    Chen, Qian
    Wang, Zilong
    Zhou, Yilin
    Chen, Jiawei
    Xiao, Dan
    Lin, Xiaodong
    INFORMATION SECURITY, ISC 2022, 2022, 13640 : 464 - 472