Asynchronous Distributed ADMM for Learning with Large-Scale and High-Dimensional Sparse Data Set

被引:2
|
作者
Wang, Dongxia [1 ]
Lei, Yongmei [1 ]
机构
[1] Shanghai Univ, Sch Comp Engn & Sci, 333 Nanchen Rd, Shanghai 200436, Peoples R China
基金
中国国家自然科学基金;
关键词
GA-ADMM; General form consensus; Bounded asynchronous; Non-convex;
D O I
10.1007/978-3-030-36405-2_27
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
The distributed alternating direction method of multipliers is an effective method to solve large-scale machine learning. At present, most distributed ADMM algorithms need to transfer the entire model parameter in the communication, which leads to high communication cost, especially when the features of model parameter is very large. In this paper, an asynchronous distributed ADMM algorithm (GA-ADMM) based on general form consensus is proposed. First, the GA-ADMM algorithm filters the information transmitted between nodes by analyzing the characteristics of high-dimensional sparse data set: only associated features, rather than all features of the model, need to be transmitted between workers and the master, thus greatly reducing the communication cost. Second, the bounded asynchronous communication protocol is used to further improve the performance of the algorithm. The convergence of the algorithm is also analyzed theoretically when the objective function is non-convex. Finally, the algorithm is tested on the cluster supercomputer "Ziqiang 4000". The experiments show that the GA-ADMM algorithm converges when appropriate parameters are selected, the GA-ADMM algorithm requires less system time to reach convergence than the AD-ADMM algorithm, and the accuracy of these two algorithms is approximate.
引用
收藏
页码:259 / 274
页数:16
相关论文
共 50 条
  • [21] Asynchronous Parallel, Sparse Approximated SVRG for High-Dimensional Machine Learning
    Shang, Fanhua
    Huang, Hua
    Fan, Jun
    Liu, Yuanyuan
    Liu, Hongying
    Liu, Jianhui
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2022, 34 (12) : 5636 - 5648
  • [22] Large-scale asynchronous distributed learning based on parameter exchanges
    Joshi, Bikash
    Iutzeler, Franck
    Amini, Massih-Reza
    INTERNATIONAL JOURNAL OF DATA SCIENCE AND ANALYTICS, 2018, 5 (04) : 223 - 232
  • [23] Scalable Collaborative Targeted Learning for Large Scale and High-Dimensional Data
    Ju, Cheng
    Gruber, Susan
    Lendle, Samuel D.
    Franklin, Jessica M.
    Wyss, Richard
    Schneeweiss, Sebastian
    van der Laan, Mark J.
    PHARMACOEPIDEMIOLOGY AND DRUG SAFETY, 2017, 26 : 529 - 530
  • [24] Asynchronous Distributed ADMM for Large-Scale Optimization-Part I: Algorithm and Convergence Analysis
    Chang, Tsung-Hui
    Hong, Mingyi
    Liao, Wei-Cheng
    Wang, Xiangfeng
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2016, 64 (12) : 3118 - 3130
  • [25] Efficient Sparse Representation for Learning With High-Dimensional Data
    Chen, Jie
    Yang, Shengxiang
    Wang, Zhu
    Mao, Hua
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2023, 34 (08) : 4208 - 4222
  • [26] Data Quality Measures and Efficient Evaluation Algorithms for Large-Scale High-Dimensional Data
    Cho, Hyeongmin
    Lee, Sangkyun
    APPLIED SCIENCES-BASEL, 2021, 11 (02): : 1 - 17
  • [27] Communication-efficient distributed estimation for high-dimensional large-scale linear regression
    Liu, Zhan
    Zhao, Xiaoluo
    Pan, Yingli
    METRIKA, 2023, 86 (04) : 455 - 485
  • [28] Distributed high-dimensional similarity search approach for large-scale wireless sensor networks
    Hu, Haifeng
    He, Jiefang
    Wu, Jianshen
    Wang, Kun
    Zhuang, Wei
    INTERNATIONAL JOURNAL OF DISTRIBUTED SENSOR NETWORKS, 2017, 13 (03):
  • [29] Communication-efficient distributed estimation for high-dimensional large-scale linear regression
    Zhan Liu
    Xiaoluo Zhao
    Yingli Pan
    Metrika, 2023, 86 : 455 - 485
  • [30] Large-Scale Distributed Sparse Class-Imbalance Learning
    Maurya, Chandresh Kumar
    Toshniwal, Durga
    INFORMATION SCIENCES, 2018, 456 : 1 - 12