Federated Learning via Disentangled Information Bottleneck

被引:5
|
作者
Uddin, Md Palash [1 ]
Xiang, Yong [1 ]
Lu, Xuequan [1 ]
Yearwood, John [2 ]
Gao, Longxiang [3 ,4 ]
机构
[1] Deakin Univ, Sch Informat Technol, Deakin Blockchain Innovat Lab, Geelong, Vic 3220, Australia
[2] Deakin Univ, Sch Informat Technol, Geelong, Vic 3220, Australia
[3] Qilu Univ Technol, Shandong Acad Sci, Jinan 250316, Shandong, Peoples R China
[4] Natl Supercomp Ctr Jinan, Shandong Comp Sci Ctr, Jinan 250014, Shandong, Peoples R China
基金
澳大利亚研究理事会;
关键词
Data models; Training; Servers; Convergence; Predictive models; Costs; Computational modeling; Federated learning; communication overhead; Index Terms; information bottleneck; mutual information; disentanglement; MUTUAL INFORMATION; NEURAL-NETWORKS;
D O I
10.1109/TSC.2022.3187962
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Existing Federated Learning (FL) algorithms generally suffer from high communication costs and data heterogeneity due to the use of conventional loss function for local model update and the equal consideration of each local model for global model aggregation. In this article, we propose a novel FL approach to address the above issues. For local model update, we propose a disentangled Information Bottleneck (IB) principle-based loss function. For global model aggregation, we suggest a model selection strategy based on Mutual Information (MI). Particularly, we design a Lagrangian-based loss function using the IB principle and "disentanglement" for maximizing MI between the ground truth and model prediction and minimizing MI between the intermediate representations. We calculate MI ratio between the ground truth and model prediction, and between the original input and ground truth to select the effective models for aggregation. We analyze the theoretical optimal cost of the loss function and manifest optimal convergence rate, and quantify the outlier robustness of the aggregation scheme. Experiments demonstrate the superiority of the proposed FL approach, in terms of testing performance and communication speedup (i.e., 3.00-14.88 times for IID MNIST, 2.5-50.75 times for non-IID MNIST, 1.87-18.40 times for IID CIFAR-10, and 1.24-2.10 times for non-IID MIMIC-III).
引用
收藏
页码:1874 / 1889
页数:16
相关论文
共 50 条
  • [1] Federated learning via reweighting information bottleneck with domain generalization
    Li, Fangyu
    Chen, Xuqiang
    Han, Zhu
    Du, Yongping
    Han, Honggui
    [J]. INFORMATION SCIENCES, 2024, 677
  • [2] Disentangled Information Bottleneck
    Pan, Ziqi
    Li Niu
    Zhang, Jianfu
    Zhang, Liqing
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 9285 - 9293
  • [3] β-CapsNet: learning disentangled representation for CapsNet by information bottleneck
    Hu, Ming-fei
    Liu, Jian-wei
    [J]. NEURAL COMPUTING & APPLICATIONS, 2023, 35 (03): : 2503 - 2525
  • [4] MVAIBNet: Multiview Disentangled Representation Learning With Information Bottleneck
    Yin, Ming
    Liu, Xin
    Gao, Junli
    Yuan, Haoliang
    Jin, Taisong
    Zhang, Shengwei
    Li, Lingling
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024,
  • [5] β-CapsNet: learning disentangled representation for CapsNet by information bottleneck
    Ming-fei Hu
    Jian-wei Liu
    [J]. Neural Computing and Applications, 2023, 35 : 2503 - 2525
  • [6] Disentangled Generation With Information Bottleneck for Enhanced Few-Shot Learning
    Dang, Zhuohang
    Luo, Minnan
    Wang, Jihong
    Jia, Chengyou
    Yan, Caixia
    Dai, Guang
    Chang, Xiaojun
    Zheng, Qinghua
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2024, 33 : 3520 - 3535
  • [7] Unsupervised MR harmonization by learning disentangled representations using information bottleneck theory
    Zuo, Lianrui
    Dewey, Blake E.
    Liu, Yihao
    He, Yufan
    Newsome, Scott D.
    Mowry, Ellen M.
    Resnick, Susan M.
    Prince, Jerry L.
    Carass, Aaron
    [J]. NEUROIMAGE, 2021, 243
  • [8] Learning Unsupervised Disentangled Capsule via Mutual Information
    Hu, MingFei
    Liu, ZeYu
    Liu, JianWei
    [J]. 2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [9] IB-GAN: Disentangled Representation Learning with Information Bottleneck Generative Adversarial Networks
    Jeon, Insu
    Lee, Wonkwang
    Pyeon, Myeongjang
    Kim, Gunhee
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 7926 - 7934
  • [10] Disentangled Speaker Representation Learning via Mutual Information Minimization
    Mun, Sung Hwan
    Han, Min Hyun
    Kim, Minchan
    Lee, Dongjune
    Kim, Nam Soo
    [J]. PROCEEDINGS OF 2022 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2022, : 89 - 96