Communication-efficient ADMM-based distributed algorithms for sparse training

被引:3
|
作者
Wang, Guozheng [1 ]
Lei, Yongmei [1 ]
Qiu, Yongwen [1 ]
Lou, Lingfei [1 ]
Li, Yixin [1 ]
机构
[1] Shanghai Univ, Sch Comp Engn & Sci, Shanghai 200444, Peoples R China
基金
中国国家自然科学基金;
关键词
ADMM; Grouped Sparse AllReduce; Two-dimensional torus topology; Synchronization algorithm;
D O I
10.1016/j.neucom.2023.126456
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In large-scale distributed machine learning (DML), the synchronization efficiency of the distributed algorithm becomes a critical factor that affects the training time of machine learning models as the computing scale increases. To address this challenge, we propose a novel algorithm called Grouped Sparse AllReduce based on the 2D-Torus topology (2D-TGSA), which enables constant transmission traffic that does not change with the number of workers. Our experimental results demonstrate that 2D-TGSA outperforms several benchmark algorithms in terms of synchronization efficiency. Moreover, we integrate the general form consistent ADMM with 2D-TGSA to develop a distributed algorithm (2D-TGSAADMM) that exhibits excellent scalability and can effectively handle large-scale distributed optimization problems. Furthermore, we enhance 2D-TGSA-ADMM by adopting the resilient adaptive penalty parameter approach, resulting in a new algorithm called 2D-TGSA-TPADMM. Our experiments on training the logistic regression model with '1-norm on the Tianhe-2 supercomputing platform demonstrate that our proposed algorithm can significantly reduce the synchronization time and training time compared to state-of-the-art methods.& COPY; 2023 Elsevier B.V. All rights reserved.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Communication-efficient Distributed Solutions to a System of Linear Equations with Laplacian Sparse Structure
    Wang, Peng
    Gao, Yuanqi
    Yu, Nanpeng
    Ren, Wei
    Lian, Jianming
    Wu, Di
    2018 IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2018, : 3367 - 3372
  • [42] An Uplink Communication-Efficient Approach to Featurewise Distributed Sparse Optimization With Differential Privacy
    Lou, Jian
    Cheung, Yiu-ming
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2021, 32 (10) : 4529 - 4543
  • [43] ADMM-SVNet: An ADMM-Based Sparse-View CT Reconstruction Network
    Wang, Sukai
    Li, Xuan
    Chen, Ping
    PHOTONICS, 2022, 9 (03)
  • [44] Fast and Communication-Efficient Algorithm for Distributed Support Vector Machine Training
    Dass, Jyotikrishna
    Sarin, Vivek
    Mahapatra, Rabi N.
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2019, 30 (05) : 1065 - 1076
  • [45] ScaleCom: Scalable Sparsified Gradient Compression for Communication-Efficient Distributed Training
    Chen, Chia-Yu
    Ni, Jiamin
    Lu, Songtao
    Cui, Xiaodong
    Chen, Pin-Yu
    Sun, Xiao
    Wang, Naigang
    Venkataramani, Swagath
    Srinivasan, Vijayalakshmi
    Zhang, Wei
    Gopalakrishnan, Kailash
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [46] Communication-Efficient Algorithms for Statistical Optimization
    Zhang, Yuchen
    Duchi, John C.
    Wainwright, Martin J.
    2012 IEEE 51ST ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2012, : 6792 - 6792
  • [47] RIS-ADMM: A RIS and ADMM-Based Passive and Sparse Sensing Method With Interference Removal
    Chen, Peng
    Chen, Zhimin
    Miao, Pu
    Chen, Yun
    IEEE COMMUNICATIONS LETTERS, 2024, 28 (04) : 867 - 871
  • [48] Communication-efficient distributed oblivious transfer
    Beimel, Amos
    Chee, Yeow Meng
    Wang, Huaxiong
    Zhang, Liang Feng
    JOURNAL OF COMPUTER AND SYSTEM SCIENCES, 2012, 78 (04) : 1142 - 1157
  • [49] Communication-Efficient Distributed Skyline Computation
    Zhang, Haoyu
    Zhang, Qin
    CIKM'17: PROCEEDINGS OF THE 2017 ACM CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2017, : 437 - 446
  • [50] Communication-Efficient Distributed Eigenspace Estimation
    Charisopoulos, Vasileios
    Benson, Austin R.
    Damle, Anil
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2021, 3 (04): : 1067 - 1092