Analyzing the Impact of Personalization on Fairness in Federated Learning for Healthcare

被引:0
|
作者
Wang, Tongnian [1 ]
Zhang, Kai [2 ]
Cai, Jiannan [3 ]
Gong, Yanmin [4 ]
Choo, Kim-Kwang Raymond [1 ]
Guo, Yuanxiong [1 ]
机构
[1] Univ Texas San Antonio, Dept Informat Syst & Cyber Secur, San Antonio, TX 78249 USA
[2] Univ Texas Hlth Sci Ctr Houston, McWilliams Sch Biomed Informat, Houston, TX 77030 USA
[3] Univ Texas San Antonio, Sch Civil & Environm Engn & Construct Management, San Antonio, TX 78249 USA
[4] Univ Texas San Antonio, Dept Elect & Comp Engn, San Antonio, TX 78249 USA
基金
美国国家科学基金会;
关键词
Health disparities; Group fairness; Federated learning; Personalization; Privacy; EXTRACTION; DISEASE;
D O I
10.1007/s41666-024-00164-7
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
As machine learning (ML) usage becomes more popular in the healthcare sector, there are also increasing concerns about potential biases and risks such as privacy. One countermeasure is to use federated learning (FL) to support collaborative learning without the need for patient data sharing across different organizations. However, the inherent heterogeneity of data distributions among participating FL parties poses challenges for exploring group fairness in FL. While personalization within FL can handle performance degradation caused by data heterogeneity, its influence on group fairness is not fully investigated. Therefore, the primary focus of this study is to rigorously assess the impact of personalized FL on group fairness in the healthcare domain, offering a comprehensive understanding of how personalized FL affects group fairness in clinical outcomes. We conduct an empirical analysis using two prominent real-world Electronic Health Records (EHR) datasets, namely eICU and MIMIC-IV. Our methodology involves a thorough comparison between personalized FL and two baselines: standalone training, where models are developed independently without FL collaboration, and standard FL, which aims to learn a global model via the FedAvg algorithm. We adopt Ditto as our personalized FL approach, which enables each client in FL to develop its own personalized model through multi-task learning. Our assessment is achieved through a series of evaluations, comparing the predictive performance (i.e., AUROC and AUPRC) and fairness gaps (i.e., EOPP, EOD, and DP) of these methods. Personalized FL demonstrates superior predictive accuracy and fairness over standalone training across both datasets. Nevertheless, in comparison with standard FL, personalized FL shows improved predictive accuracy but does not consistently offer better fairness outcomes. For instance, in the 24-h in-hospital mortality prediction task, personalized FL achieves an average EOD of 27.4% across racial groups in the eICU dataset and 47.8% in MIMIC-IV. In comparison, standard FL records a better EOD of 26.2% for eICU and 42.0% for MIMIC-IV, while standalone training yields significantly worse EOD of 69.4% and 54.7% on these datasets, respectively. Our analysis reveals that personalized FL has the potential to enhance fairness in comparison to standalone training, yet it does not consistently ensure fairness improvements compared to standard FL. Our findings also show that while personalization can improve fairness for more biased hospitals (i.e., hospitals having larger fairness gaps in standalone training), it can exacerbate fairness issues for less biased ones. These insights suggest that the integration of personalized FL with additional strategic designs could be key to simultaneously boosting prediction accuracy and reducing fairness disparities. The findings and opportunities outlined in this paper can inform the research agenda for future studies, to overcome the limitations and further advance health equity research.
引用
收藏
页码:181 / 205
页数:25
相关论文
共 50 条
  • [21] Ditto: Fair and Robust Federated Learning Through Personalization
    Li, Tian
    Hu, Shengyuan
    Beirami, Ahmad
    Smith, Virginia
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [22] On Privacy and Personalization in Cross-Silo Federated Learning
    Liu, Ziyu
    Hu, Shengyuan
    Wu, Zhiwei Steven
    Smith, Virginia
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [23] Improving Global Generalization and Local Personalization for Federated Learning
    Meng, Lei
    Qi, Zhuang
    Wu, Lei
    Du, Xiaoyu
    Li, Zhaochuan
    Cui, Lizhen
    Meng, Xiangxu
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (01) : 76 - 87
  • [24] Federated Learning for Healthcare Applications
    Chaddad, Ahmad
    Wu, Yihang
    Desrosiers, Christian
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (05): : 7339 - 7358
  • [25] Federated Learning for Healthcare Informatics
    Jie Xu
    Benjamin S. Glicksberg
    Chang Su
    Peter Walker
    Jiang Bian
    Fei Wang
    Journal of Healthcare Informatics Research, 2021, 5 : 1 - 19
  • [26] Towards Fairness-Aware Federated Learning
    Shi, Yuxin
    Yu, Han
    Leung, Cyril
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (09) : 11922 - 11938
  • [27] FairFed: Enabling Group Fairness in Federated Learning
    Ezzeldin, Yahya H.
    Yan, Shen
    He, Chaoyang
    Ferrara, Emilio
    Avestimehr, Salman
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 6, 2023, : 7494 - 7502
  • [28] Federated Learning for Healthcare Informatics
    Xu, Jie
    Glicksberg, Benjamin S.
    Su, Chang
    Walker, Peter
    Bian, Jiang
    Wang, Fei
    JOURNAL OF HEALTHCARE INFORMATICS RESEARCH, 2021, 5 (01) : 1 - 19
  • [29] Fairness and privacy preserving in federated learning: A survey
    Rafi, Taki Hasan
    Noor, Faiza Anan
    Hussain, Tahmid
    Chae, Dong-Kyu
    INFORMATION FUSION, 2024, 105
  • [30] Privacy and Fairness in Federated Learning: On the Perspective of Tradeoff
    Chen, Huiqiang
    Zhu, Tianqing
    Zhang, Tao
    Zhou, Wanlei
    Yu, Philip S.
    ACM COMPUTING SURVEYS, 2024, 56 (02)