Asynchronous Distributed ADMM for Learning with Large-Scale and High-Dimensional Sparse Data Set

被引:2
|
作者
Wang, Dongxia [1 ]
Lei, Yongmei [1 ]
机构
[1] Shanghai Univ, Sch Comp Engn & Sci, 333 Nanchen Rd, Shanghai 200436, Peoples R China
基金
中国国家自然科学基金;
关键词
GA-ADMM; General form consensus; Bounded asynchronous; Non-convex;
D O I
10.1007/978-3-030-36405-2_27
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
The distributed alternating direction method of multipliers is an effective method to solve large-scale machine learning. At present, most distributed ADMM algorithms need to transfer the entire model parameter in the communication, which leads to high communication cost, especially when the features of model parameter is very large. In this paper, an asynchronous distributed ADMM algorithm (GA-ADMM) based on general form consensus is proposed. First, the GA-ADMM algorithm filters the information transmitted between nodes by analyzing the characteristics of high-dimensional sparse data set: only associated features, rather than all features of the model, need to be transmitted between workers and the master, thus greatly reducing the communication cost. Second, the bounded asynchronous communication protocol is used to further improve the performance of the algorithm. The convergence of the algorithm is also analyzed theoretically when the objective function is non-convex. Finally, the algorithm is tested on the cluster supercomputer "Ziqiang 4000". The experiments show that the GA-ADMM algorithm converges when appropriate parameters are selected, the GA-ADMM algorithm requires less system time to reach convergence than the AD-ADMM algorithm, and the accuracy of these two algorithms is approximate.
引用
收藏
页码:259 / 274
页数:16
相关论文
共 50 条
  • [1] A Supervised Learning Model for High-Dimensional and Large-Scale Data
    Peng, Chong
    Cheng, Jie
    Cheng, Qiang
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2017, 8 (02)
  • [2] Efficient distributed optimization for large-scale high-dimensional sparse penalized Huber regression
    Pan, Yingli
    Xu, Kaidong
    Wei, Sha
    Wang, Xiaojuan
    Liu, Zhan
    COMMUNICATIONS IN STATISTICS-SIMULATION AND COMPUTATION, 2024, 53 (07) : 3106 - 3125
  • [3] Sparse Learning for Large-Scale and High-Dimensional Data: A Randomized Convex-Concave Optimization Approach
    Zhang, Lijun
    Yang, Tianbao
    Jin, Rong
    Zhou, Zhi-Hua
    ALGORITHMIC LEARNING THEORY, (ALT 2016), 2016, 9925 : 83 - 97
  • [4] Visualizing Large-scale and High-dimensional Data
    Tang, Jian
    Liu, Jingzhou
    Zhang, Ming
    Mei, Qiaozhu
    PROCEEDINGS OF THE 25TH INTERNATIONAL CONFERENCE ON WORLD WIDE WEB (WWW'16), 2016, : 287 - 297
  • [5] Distributed Methods for High-dimensional and Large-scale Tensor Factorization
    Shin, Kijung
    Kang, U.
    2014 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2014, : 989 - 994
  • [6] High-dimensional MRI data analysis using a large-scale manifold learning approach
    Loc Tran
    Debrup Banerjee
    Jihong Wang
    Ashok J. Kumar
    Frederic McKenzie
    Yaohang Li
    Jiang Li
    Machine Vision and Applications, 2013, 24 : 995 - 1014
  • [7] High-dimensional MRI data analysis using a large-scale manifold learning approach
    Tran, Loc
    Banerjee, Debrup
    Wang, Jihong
    Kumar, Ashok J.
    McKenzie, Frederic
    Li, Yaohang
    Li, Jiang
    MACHINE VISION AND APPLICATIONS, 2013, 24 (05) : 995 - 1014
  • [8] Fast Low-rank Metric Learning for Large-scale and High-dimensional Data
    Liu, Han
    Han, Zhizhong
    Liu, Yu-Shen
    Gu, Ming
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [9] RECURSIVE REDUCTION NET FOR LARGE-SCALE HIGH-DIMENSIONAL DATA
    Ke, Tsung-Wei
    Liu, Tyng-Luh
    2016 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2016, : 1903 - 1907
  • [10] Data Independent Method of Constructing Distributed LSH for Large-Scale Dynamic High-Dimensional Indexing
    Gu, Xiaoguang
    Zhang, Lei
    Zhang, Dongming
    Zhang, Yongdong
    Li, Jintao
    Bao, Ning
    2012 IEEE 14TH INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE COMPUTING AND COMMUNICATIONS & 2012 IEEE 9TH INTERNATIONAL CONFERENCE ON EMBEDDED SOFTWARE AND SYSTEMS (HPCC-ICESS), 2012, : 564 - 571