HCFL: A High Compression Approach for Communication-Efficient Federated Learning in Very Large Scale IoT Networks

被引:7
|
作者
Nguyen, Minh-Duong [1 ]
Lee, Sang-Min [1 ]
Pham, Quoc-Viet [2 ]
Hoang, Dinh Thai [3 ]
Nguyen, Diep N. [3 ]
Hwang, Won-Joo [4 ]
机构
[1] Pusan Natl Univ, Dept Informat Convergence Engn, Pusan 46241, South Korea
[2] Pusan Natl Univ, Korean Southeast Ctr Ind Revolut Leader Educ 4, Pusan 46241, South Korea
[3] Univ Technol Sydney, Sch Elect & Data Engn, Sydney, NSW 2007, Australia
[4] Pusan Natl Univ, Dept Biomed Convergence Engn, Yangsan 50612, South Korea
基金
澳大利亚研究理事会; 新加坡国家研究基金会;
关键词
Autoencoder; communication efficiency; data compression; deep learning; distributed learning; federated learning; internet-of-things; machine type communication;
D O I
10.1109/TMC.2022.3190510
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Federated learning (FL) is a new artificial intelligence concept that enables Internet-of-Things (IoT) devices to learn a collaborative model without sending the raw data to centralized nodes for processing. Despite numerous advantages, low computing resources at IoT devices and high communication costs for exchanging model parameters make applications of FL in massive IoT networks very limited. In this work, we develop a novel compression scheme for FL, called high-compression federated learning (HCFL) , for very large scale IoT networks. HCFL can reduce the data load for FL processes without changing their structure and hyperparameters. In this way, we not only can significantly reduce communication costs, but also make intensive learning processes more adaptable on low-computing resource IoT devices. Furthermore, we investigate a relationship between the number of IoT devices and the convergence level of the FL model and thereby better assess the quality of the FL process. We demonstrate our HCFL scheme in both simulations and mathematical analyses. Our proposed theoretical research can be used as a minimum level of satisfaction, proving that the FL process can achieve good performance when a determined configuration is met. Therefore, we show that HCFL is applicable in any FL-integrated networks with numerous IoT devices.
引用
下载
收藏
页码:6495 / 6507
页数:13
相关论文
共 50 条
  • [41] FedHe: Heterogeneous Models and Communication-Efficient Federated Learning
    Chan, Yun Hin
    Ngai, Edith C. H.
    2021 17TH INTERNATIONAL CONFERENCE ON MOBILITY, SENSING AND NETWORKING (MSN 2021), 2021, : 207 - 214
  • [42] Communication-Efficient Robust Federated Learning with Noisy Labels
    Li, Junyi
    Pei, Jian
    Huang, Heng
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 914 - 924
  • [43] FedADP: Communication-Efficient by Model Pruning for Federated Learning
    Liu, Haiyang
    Shi, Yuliang
    Su, Zhiyuan
    Zhang, Kun
    Wang, Xinjun
    Yan, Zhongmin
    Kong, Fanyu
    IEEE CONFERENCE ON GLOBAL COMMUNICATIONS, GLOBECOM, 2023, : 3093 - 3098
  • [44] Communication-Efficient Consensus Mechanism for Federated Reinforcement Learning
    Xu, Xing
    Li, Rongpeng
    Zhao, Zhifeng
    Zhang, Honggang
    IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC 2022), 2022, : 80 - 85
  • [45] Communication-efficient Federated Learning with Cooperative Filter Selection
    Yang, Zhao
    Sun, Qingshuang
    2022 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS 22), 2022, : 2172 - 2176
  • [46] Communication-Efficient Federated Learning with Adaptive Consensus ADMM
    He, Siyi
    Zheng, Jiali
    Feng, Minyu
    Chen, Yixin
    APPLIED SCIENCES-BASEL, 2023, 13 (09):
  • [47] Communication-Efficient Federated Learning With Gradual Layer Freezing
    Malan, Erich
    Peluso, Valentino
    Calimera, Andrea
    Macii, Enrico
    IEEE EMBEDDED SYSTEMS LETTERS, 2023, 15 (01) : 25 - 28
  • [48] Communication-Efficient Federated Learning Based on Compressed Sensing
    Li, Chengxi
    Li, Gang
    Varshney, Pramod K.
    IEEE INTERNET OF THINGS JOURNAL, 2021, 8 (20) : 15531 - 15541
  • [49] On the Convergence of Communication-Efficient Local SGD for Federated Learning
    Gao, Hongchang
    Xu, An
    Huang, Heng
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 7510 - 7518
  • [50] IoT Device Friendly and Communication-Efficient Federated Learning via Joint Model Pruning and Quantization
    Prakash, Pavana
    Ding, Jiahao
    Chen, Rui
    Qin, Xiaoqi
    Shu, Minglei
    Cui, Qimei
    Guo, Yuanxiong
    Pan, Miao
    IEEE INTERNET OF THINGS JOURNAL, 2022, 9 (15): : 13638 - 13650