Model Sparsification for Communication-Efficient Multi-Party Learning via Contrastive Distillation in Image Classification

被引:3
|
作者
Feng, Kai-Yuan [1 ,2 ]
Gong, Maoguo [1 ,3 ]
Pan, Ke [1 ,5 ]
Zhao, Hongyu [1 ,3 ]
Wu, Yue [1 ,4 ]
Sheng, Kai [1 ,2 ]
机构
[1] Xidian Univ, Key Lab Collaborat Intelligence Syst, Minist Educ, Xian 710071, Peoples R China
[2] Xidian Univ, Acad Adv Interdisciplinary Res, Xian 710071, Peoples R China
[3] Xidian Univ, Sch Elect Engn, Xian 710071, Peoples R China
[4] Xidian Univ, Sch Comp Sci & Technol, Xian 710071, Peoples R China
[5] Xidian Univ, Sch Cyber Engn, Xian 710071, Peoples R China
基金
中国国家自然科学基金;
关键词
Data models; Computational modeling; Servers; Adaptation models; Training; Feature extraction; Performance evaluation; Multi-party learning; model sparsification; contrastive distillation; efficient communication; NEURAL-NETWORKS;
D O I
10.1109/TETCI.2023.3268713
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-party learning allows all parties to train a joint model under legal and practical constraints without private data transmission. Related research can perform multi-party learning tasks on homogeneous data through deep networks. However, due to the heterogeneity of data from different parties and the limitation of computational resources and costs, traditional approaches may affect the effectiveness of multi-party learning, and cannot provide a personalized network for each party. In addition, to reduce the computational cost and communication bandwidth of local models, there are still challenges in building an adaptive model from the private data of different parties. To address these challenges, we aim to apply a model sparsification strategy in multi-party learning. Model sparsification can not only reduce the computational overhead in local edge devices and the cost of communication and interaction between multi-party models. It can also develop privatized and personalized networks based on the heterogeneity of local data. We use the contrastive distillation method during training to reduce the distance between local and global models. In addition, we maintain the performance of the aggregation model from heterogeneous data. In brief, we developed an adaptive multi-party learning framework based on contrastive distillation, which can significantly reduce the communication cost in the learning process, improve the effectiveness of the aggregation model for local heterogeneous and unbalanced data, and make it easy to deploy in the limited edge devices. Finally, to verify the effectiveness of this framework, we experimented with the Fshion-MNIST, Cifar-10, and Cifar-100 datasets in different scenarios to verify the effectiveness of this framework.
引用
收藏
页码:150 / 163
页数:14
相关论文
共 50 条
  • [1] ATOMO: Communication-efficient Learning via Atomic Sparsification
    Wang, Hongyi
    Sievert, Scott
    Charles, Zachary
    Liu, Shengchao
    Wright, Stephen
    Papailiopoulos, Dimitris
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [2] SCALR: Communication-Efficient Secure Multi-Party Logistic Regression
    Lu, Xingyu
    Sami, Hasin Us
    Guler, Basak
    IEEE TRANSACTIONS ON COMMUNICATIONS, 2024, 72 (01) : 162 - 178
  • [3] Communication-efficient federated learning via knowledge distillation
    Wu, Chuhan
    Wu, Fangzhao
    Lyu, Lingjuan
    Huang, Yongfeng
    Xie, Xing
    NATURE COMMUNICATIONS, 2022, 13 (01)
  • [4] Communication-Efficient Federated Learning for Multi-Institutional Medical Image Classification
    Zhou, Shuang
    Landman, Bennett A.
    Huo, Yuankai
    Gokhale, Aniruddha
    MEDICAL IMAGING 2022: IMAGING INFORMATICS FOR HEALTHCARE, RESEARCH, AND APPLICATIONS, 2022, 12037
  • [5] Time-Correlated Sparsification for Communication-Efficient Federated Learning
    Ozfatura, Emre
    Ozfatura, Kerem
    Gunduz, Deniz
    2021 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2021, : 461 - 466
  • [6] ACFL: Communication-Efficient adversarial contrastive federated learning for medical image segmentation
    Liang, Zhiwei
    Zhao, Kui
    Liang, Gang
    Wu, Yifei
    Guo, Jinxi
    KNOWLEDGE-BASED SYSTEMS, 2024, 304
  • [7] Communication-Efficient Decentralized Learning with Sparsification and Adaptive Peer Selection
    Tang, Zhenheng
    Shi, Shaohuai
    Chu, Xiaowen
    2020 IEEE 40TH INTERNATIONAL CONFERENCE ON DISTRIBUTED COMPUTING SYSTEMS (ICDCS), 2020, : 1207 - 1208
  • [8] Communication-Efficient Distributed Deep Learning with Merged Gradient Sparsification on GPUs
    Shi, Shaohuai
    Wang, Qiang
    Chu, Xiaowen
    Li, Bo
    Qin, Yang
    Liu, Ruihao
    Zhao, Xinxiao
    IEEE INFOCOM 2020 - IEEE CONFERENCE ON COMPUTER COMMUNICATIONS, 2020, : 406 - 415
  • [9] Communication-efficient clustered federated learning via model distance
    Zhang, Mao
    Zhang, Tie
    Cheng, Yifei
    Bao, Changcun
    Cao, Haoyu
    Jiang, Deqiang
    Xu, Linli
    MACHINE LEARNING, 2024, 113 (06) : 3869 - 3888
  • [10] Communication-efficient clustered federated learning via model distance
    Mao Zhang
    Tie Zhang
    Yifei Cheng
    Changcun Bao
    Haoyu Cao
    Deqiang Jiang
    Linli Xu
    Machine Learning, 2024, 113 : 3869 - 3888