Personalized Federated Learning with Theoretical Guarantees: A Model-Agnostic Meta-Learning Approach

被引:0
|
作者
Fallah, Alireza [1 ]
Mokhtari, Aryan [2 ]
Ozdaglar, Asuman [1 ]
机构
[1] MIT, EECS Dept, Cambridge, MA 02139 USA
[2] Univ Texas Austin, ECE Dept, Austin, TX 78712 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In Federated Learning, we aim to train models across multiple computing units (users), while users can only communicate with a common central server, without exchanging their data samples. This mechanism exploits the computational power of all users and allows users to obtain a richer model as their models are trained over a larger set of data points. However, this scheme only develops a common output for all the users, and, therefore, it does not adapt the model to each user. This is an important missing feature, especially given the heterogeneity of the underlying data distribution for various users. In this paper, we study a personalized variant of the federated learning in which our goal is to find an initial shared model that current or new users can easily adapt to their local dataset by performing one or a few steps of gradient descent with respect to their own data. This approach keeps all the benefits of the federated learning architecture, and, by structure, leads to a more personalized model for each user. We show this problem can be studied within the Model-Agnostic Meta-Learning (MAML) framework. Inspired by this connection, we study a personalized variant of the well-known Federated Averaging algorithm and evaluate its performance in terms of gradient norm for non-convex loss functions. Further, we characterize how this performance is affected by the closeness of underlying distributions of user data, measured in terms of distribution distances such as Total Variation and 1-Wasserstein metric.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Memory-Based Optimization Methods for Model-Agnostic Meta-Learning and Personalized Federated Learning
    Wang, Bokun
    Yuan, Zhuoning
    Ying, Yiming
    Yang, Tianbao
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [2] Probabilistic Model-Agnostic Meta-Learning
    Finn, Chelsea
    Xu, Kelvin
    Levine, Sergey
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [3] Bayesian Model-Agnostic Meta-Learning
    Yoon, Jaesik
    Kim, Taesup
    Dia, Ousmane
    Kim, Sungwoong
    Bengio, Yoshua
    Ahn, Sungjin
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [4] Meta weight learning via model-agnostic meta-learning
    Xu, Zhixiong
    Chen, Xiliang
    Tang, Wei
    Lai, Jun
    Cao, Lei
    [J]. NEUROCOMPUTING, 2021, 432 : 124 - 132
  • [5] Knowledge Distillation for Model-Agnostic Meta-Learning
    Zhang, Min
    Wang, Donglin
    Gai, Sibo
    [J]. ECAI 2020: 24TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, 325 : 1355 - 1362
  • [6] Theoretical Convergence of Multi-Step Model-Agnostic Meta-Learning
    Ji, Kaiyi
    Yang, Junjie
    Liang, Yingbin
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [7] Theoretical Convergence of Multi-Step Model-Agnostic Meta-Learning
    Ji, Kaiyi
    Yang, Junjie
    Liang, Yingbin
    [J]. Journal of Machine Learning Research, 2022, 23
  • [8] Model-Agnostic Federated Learning
    Mittone, Gianluca
    Riviera, Walter
    Colonnelli, Iacopo
    Birke, Robert
    Aldinucci, Marco
    [J]. EURO-PAR 2023: PARALLEL PROCESSING, 2023, 14100 : 383 - 396
  • [9] Combining Model-Agnostic Meta-Learning and Transfer Learning for Regression
    Satrya, Wahyu Fadli
    Yun, Ji-Hoon
    [J]. SENSORS, 2023, 23 (02)
  • [10] Task-Robust Model-Agnostic Meta-Learning
    Collins, Liam
    Mokhtari, Aryan
    Shakkottai, Sanjay
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33