SNAP: A Communication Efficient Distributed Machine Learning Framework for Edge Computing

被引:6
|
作者
Zhao, Yangming [1 ]
Fan, Jingyuan [1 ]
Su, Lu [1 ]
Song, Tongyu [2 ]
Wang, Sheng [2 ]
Qiao, Chunming [1 ]
机构
[1] Univ Buffalo, Dept Comp Sci & Engn, Buffalo, NY USA
[2] Univ Elect Sci & Technol China, Sch Informat & Commun Engn, Chengdu, Peoples R China
关键词
NEURAL-NETWORKS;
D O I
10.1109/ICDCS47774.2020.00072
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
More and more applications learn from the data collected by the edge devices. Conventional learning methods, such as gathering all the raw data to train an ultimate model in a centralized way, or training a target model in a distributed manner under the parameter server framework, suffer a high communication cost. In this paper, we design Select Neighbors and Parameters (SNAP), a communication efficient distributed machine learning framework, to mitigate the communication cost. A distinct feature of SNAP is that the edge servers act as peers to each other. Specifically, in SNAP, every edge server hosts a copy of the global model, trains it with the local data, and periodically updates the local parameters based on the weighted sum of the parameters from its neighbors (i.e., peers) only (i.e., without pulling the parameters from all other edge servers). Different from most of the previous works on consensus optimization in which the weight matrix to update parameter values is predefined, we propose a scheme to optimize the weight matrix based on the network topology, and hence the convergence rate can be improved. Another key idea in SNAP is that only the parameters which have been changed significantly since the last iteration will be sent to the neighbors. Both theoretical analysis and simulations show that SNAP can achieve the same accuracy performance as the centralized training method. Compared to the state-of-the-art communication-aware distributed learning scheme TernGrad, SNAP incurs a significantly lower (99.6% lower) communication cost.
引用
收藏
页码:584 / 594
页数:11
相关论文
共 50 条
  • [1] UbiNN: A Communication Efficient Framework for Distributed Machine Learning in Edge Computing
    Li, Ke
    Chen, Kexun
    Luo, Shouxi
    Zhang, Honghao
    Fan, Pingzhi
    [J]. IEEE TRANSACTIONS ON NETWORK SCIENCE AND ENGINEERING, 2023, 10 (06): : 3368 - 3383
  • [2] GADMM: Fast and Communication Efficient Framework for Distributed Machine Learning
    Elgabli, Anis
    Park, Jihong
    Bedi, Amrit S.
    Bennis, Mehdi
    Aggarwal, Vaneet
    [J]. Journal of Machine Learning Research, 2020, 21
  • [3] GADMM: Fast and Communication Efficient Framework for Distributed Machine Learning
    Elgabli, Anis
    Park, Jihong
    Bedi, Amrit S.
    Bennis, Mehdi
    Aggarwal, Vaneet
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [4] Edge Computing Solutions for Distributed Machine Learning
    Marozzo, Fabrizio
    Orsino, Alessio
    Talia, Domenico
    Trunfio, Paolo
    [J]. 2022 IEEE INTL CONF ON DEPENDABLE, AUTONOMIC AND SECURE COMPUTING, INTL CONF ON PERVASIVE INTELLIGENCE AND COMPUTING, INTL CONF ON CLOUD AND BIG DATA COMPUTING, INTL CONF ON CYBER SCIENCE AND TECHNOLOGY CONGRESS (DASC/PICOM/CBDCOM/CYBERSCITECH), 2022, : 1148 - 1155
  • [5] An Edge Computing Marketplace for Distributed Machine Learning
    Yerabolu, Susham
    Gomena, Samuel
    Aryafar, Ehsan
    Joe-Wong, Carlee
    [J]. PROCEEDINGS OF THE 2019 ACM SIGCOMM CONFERENCE POSTERS AND DEMOS (SIGCOMM '19), 2019, : 36 - 38
  • [6] Efficient Distributed Swarm Learning for Edge Computing
    Fan, Xin
    Wang, Yue
    Huo, Yan
    Tian, Zhi
    [J]. ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 3627 - 3632
  • [7] Communication Efficient Framework for Decentralized Machine Learning
    Elgabli, Anis
    Park, Jihong
    Bedi, Amrit S.
    Bennis, Mehdi
    Aggarwal, Vaneet
    [J]. 2020 54TH ANNUAL CONFERENCE ON INFORMATION SCIENCES AND SYSTEMS (CISS), 2020, : 47 - 51
  • [8] A Systematic Literature Review on Distributed Machine Learning in Edge Computing
    Poncinelli Filho, Carlos
    Marques Jr, Elias
    Chang, Victor
    dos Santos, Leonardo
    Bernardini, Flavia
    Pires, Paulo F.
    Ochi, Luiz
    Delicato, Flavia C.
    [J]. SENSORS, 2022, 22 (07)
  • [9] Distributed Machine Learning for Multiuser Mobile Edge Computing Systems
    Guo, Yinghao
    Zhao, Rui
    Lai, Shiwei
    Fan, Lisheng
    Lei, Xianfu
    Karagiannidis, George K.
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2022, 16 (03) : 460 - 473
  • [10] Coded Computing for Distributed Machine Learning in Wireless Edge Network
    Dhakal, Sagar
    Prakash, Saurav
    Yona, Yair
    Talwar, Shilpa
    Himayat, Nageen
    [J]. 2019 IEEE 90TH VEHICULAR TECHNOLOGY CONFERENCE (VTC2019-FALL), 2019,