Communication-efficient ADMM-based distributed algorithms for sparse training

被引:3
|
作者
Wang, Guozheng [1 ]
Lei, Yongmei [1 ]
Qiu, Yongwen [1 ]
Lou, Lingfei [1 ]
Li, Yixin [1 ]
机构
[1] Shanghai Univ, Sch Comp Engn & Sci, Shanghai 200444, Peoples R China
基金
中国国家自然科学基金;
关键词
ADMM; Grouped Sparse AllReduce; Two-dimensional torus topology; Synchronization algorithm;
D O I
10.1016/j.neucom.2023.126456
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In large-scale distributed machine learning (DML), the synchronization efficiency of the distributed algorithm becomes a critical factor that affects the training time of machine learning models as the computing scale increases. To address this challenge, we propose a novel algorithm called Grouped Sparse AllReduce based on the 2D-Torus topology (2D-TGSA), which enables constant transmission traffic that does not change with the number of workers. Our experimental results demonstrate that 2D-TGSA outperforms several benchmark algorithms in terms of synchronization efficiency. Moreover, we integrate the general form consistent ADMM with 2D-TGSA to develop a distributed algorithm (2D-TGSAADMM) that exhibits excellent scalability and can effectively handle large-scale distributed optimization problems. Furthermore, we enhance 2D-TGSA-ADMM by adopting the resilient adaptive penalty parameter approach, resulting in a new algorithm called 2D-TGSA-TPADMM. Our experiments on training the logistic regression model with '1-norm on the Tianhe-2 supercomputing platform demonstrate that our proposed algorithm can significantly reduce the synchronization time and training time compared to state-of-the-art methods.& COPY; 2023 Elsevier B.V. All rights reserved.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Efficient ADMM-Based Algorithms for Convolutional Sparse Coding
    Veshki, Farshad
    Vorobyov, Sergiy
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 389 - 393
  • [2] Local Stochastic ADMM for Communication-Efficient Distributed Learning
    ben Issaid, Chaouki
    Elgabli, Anis
    Bennis, Mehdi
    2022 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2022, : 1880 - 1885
  • [3] Improving the Privacy and Accuracy of ADMM-Based Distributed Algorithms
    Zhang, Xueru
    Khalili, Mohammad Mandi
    Liu, Mingyan
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [4] Energy Efficient Distributed Volterra Modeling Approach with ADMM-Based Sparse Signal Recovery
    Gupta, Saurav
    Sahoo, Ajit Kumar
    Sahoo, Upendra Kumar
    WIRELESS PERSONAL COMMUNICATIONS, 2021, 119 (03) : 2755 - 2773
  • [5] Consensus ADMM-Based Distributed Simultaneous Imaging & Communication
    Mehrotra, Nishant
    Sabharwal, Ashutosh
    Uribe, Cesar A.
    IFAC PAPERSONLINE, 2022, 55 (13): : 31 - 36
  • [6] Energy Efficient Distributed Volterra Modeling Approach with ADMM-Based Sparse Signal Recovery
    Saurav Gupta
    Ajit Kumar Sahoo
    Upendra Kumar Sahoo
    Wireless Personal Communications, 2021, 119 : 2755 - 2773
  • [7] ADMM-Based Sparse Distributed Learning for Stochastic Configuration Networks
    Zhou, Yujun
    Ai, Wu
    2020 CHINESE AUTOMATION CONGRESS (CAC 2020), 2020, : 4354 - 4358
  • [8] Learning-based Adaptive Quantization for Communication-efficient Distributed Optimization with ADMM
    Nghiem, Truong X.
    Duarte, Aldo
    Wei, Shuangqing
    2020 54TH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS, 2020, : 37 - 41
  • [9] Communication-Efficient Distributed Demand Response: A Randomized ADMM Approach
    Tsai, Shin-Ching
    Tseng, Yi-Hen
    Chang, Tsung-Hui
    IEEE TRANSACTIONS ON SMART GRID, 2017, 8 (03) : 1085 - 1095
  • [10] More communication-efficient distributed sparse learning
    Zhou, Xingcai
    Yang, Guang
    INFORMATION SCIENCES, 2024, 668