Function Placement for In-network Federated Learning

被引:0
|
作者
Yellas, Nour-El-Houda [1 ,2 ]
Addis, Bernardetta [3 ]
Boumerdassi, Selma [1 ]
Riggio, Roberto [4 ]
Secci, Stefano [1 ]
机构
[1] Cnam, Paris, France
[2] Orange, Chatillon, France
[3] Univ Lorraine, CNRS, LORIA, Nancy, France
[4] Polytech Univ Marche, Ancona, Italy
基金
欧盟地平线“2020”;
关键词
Federated learning; Artificial intelligence functions; Placement; EDGE INTELLIGENCE; CLIENT SELECTION; FRAMEWORK;
D O I
10.1016/j.comnet.2024.110900
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning (FL), particularly when data is distributed across multiple clients, helps reducing the learning time by avoiding training on a massive pile-up of data. Nonetheless, low computation capacities or poor network conditions can worsen the convergence time, therefore decreasing accuracy and learning performance. In this paper, we propose a framework to deploy FL clients in a network, while compensating end-to-end time variation due to heterogeneous network setting. We present a new distributed learning control scheme, named In-network Federated Learning Control (IFLC), to support the operations of distributed federated learning functions in geographically distributed networks, and designed to mitigate the stragglers with lower deployment costs. IFLC adapts the allocation of distributed hardware accelerators to modulate the importance of local training latency in the end-to-end delay of federated learning applications, considering both deterministic and stochastic delay scenarios. By extensive simulation on realistic instances of an in-network anomaly detection application, we show that the absence of hardware accelerators can strongly impair the learning efficiency. Additionally, we show that providing hardware accelerators at only 50% of the nodes, can reduce the number of stragglers by at least 50% and up to 100% with respect to a baseline FIRST-FIT algorithm, while also lowering the deployment cost by up to 30% with respect to the case without hardware accelerators. Finally, we explore the effect of topology changes on IFLC across both hierarchical and flat topologies.
引用
收藏
页数:18
相关论文
共 50 条
  • [31] Caching and Placement for In-Network Caching in Device-to-Device Communications
    Soleimani, Somayeh
    Tao, Xiaofeng
    WIRELESS COMMUNICATIONS & MOBILE COMPUTING, 2018,
  • [32] Virtual Network Function Placement Optimization With Deep Reinforcement Learning
    Solozabal, Ruben
    Ceberio, Josu
    Sanchoyerto, Aitor
    Zabala, Luis
    Blanco, Bego
    Liberal, Fidel
    IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2020, 38 (02) : 292 - 303
  • [33] Cooperative Learning for Online In-Network Performance Monitoring
    Joseph, S. B.
    Loo, H. R.
    Ismail, I.
    Marsono, M. N.
    2015 IEEE 12TH MALAYSIA INTERNATIONAL CONFERENCE ON COMMUNICATIONS (MICC), 2015, : 213 - 218
  • [34] Cooperative Learning for Distributed In-Network Traffic Classification
    Joseph, S. B.
    Loo, H. R.
    Ismail, I.
    Andromeda, T.
    Marsono, M. N.
    IAES INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING, COMPUTER SCIENCE AND INFORMATICS, 2017, 190
  • [35] ATP: In-network Aggregation for Multi -tenant Learning
    Lao, ChonLam
    Le, Yanfang
    Mahajan, Kshiteej
    Chen, Yixi
    Wu, Wenfei
    Akella, Aditya
    Swift, Michael
    PROCEEDINGS OF THE 18TH USENIX SYMPOSIUM ON NETWORKED SYSTEM DESIGN AND IMPLEMENTATION, 2021, : 741 - 762
  • [36] Fast Learning Enabled by In-Network Drift Detection
    Xavier, Bruno Missi
    Martinello, Magnos
    Trois, Celio
    Mello, Brenno
    Rios, Ricardo
    PROCEEDINGS OF THE 8TH ASIA-PACIFIC WORKSHOP ON NETWORKING, APNET 2024, 2024, : 129 - 134
  • [37] In-Network Learning: Distributed Training and Inference in Networks
    Moldoveanu, Matei
    Zaidi, Abdellatif
    ENTROPY, 2023, 25 (06)
  • [38] On Learning Suitable Caching Policies for In-Network Caching
    Pires, Stefani
    Ribeiro, Adriana
    Sampaio, Leobino N.
    IEEE Transactions on Machine Learning in Communications and Networking, 2024, 2 : 1076 - 1092
  • [39] Scaling Distributed Machine Learning with In-Network Aggregation
    Sapio, Amedeo
    Canini, Marco
    Ho, Chen-Yu
    Nelson, Jacob
    Kalnis, Panos
    Kim, Changhoon
    Krishnamurthy, Arvind
    Moshref, Masoud
    Ports, Dan R. K.
    Richtarik, Peter
    PROCEEDINGS OF THE 18TH USENIX SYMPOSIUM ON NETWORKED SYSTEM DESIGN AND IMPLEMENTATION, 2021, : 785 - 808
  • [40] In-Network Machine Learning Using Programmable Network Devices: A Survey
    Zheng, Changgang
    Hong, Xinpeng
    Ding, Damu
    Vargaftik, Shay
    Ben-Itzhak, Yaniv
    Zilberman, Noa
    IEEE COMMUNICATIONS SURVEYS AND TUTORIALS, 2024, 26 (02): : 1171 - 1200