Survey on Parallel and Distributed Optimization Algorithms for Scalable Machine Learning

被引:0
|
作者
Kang, Liang-Yi [1 ,2 ]
Wang, Jian-Fei [1 ,2 ]
Liu, Jie [1 ,3 ]
Ye, Dan [1 ]
机构
[1] Technology Center of Software Engineering, Institute of Software, The Chinese Academy of Sciences, Beijing,100190, China
[2] University of Chinese Academy of Sciences, Beijing,100190, China
[3] State Key Laboratory of Computer Science, Institute of Software, The Chinese Academy of Sciences, Beijing,100190, China
来源
Ruan Jian Xue Bao/Journal of Software | 2018年 / 29卷 / 01期
基金
中国国家自然科学基金;
关键词
Gradient methods - Optimization - Distributed computer systems - Learning algorithms;
D O I
10.13328/j.cnki.jos.005376
中图分类号
学科分类号
摘要
Machine learning problems can be viewed as optimization-centric programs, and the optimization algorithm is an important tool to solve the objective function. In the era of big data, in order to speed up the training process, it is essential to design parallel and distributed optimization algorithms by multi-core computing and distributed computing technologies. In recent years, there are a lot of research works in this field, and some algorithms have been widely applied on machine learning platforms. In this paper, five common optimization algorithms, including gradient descent algorithm, second order optimization algorithm, proximal gradient algorithm, coordinate descent algorithm and alternating direction method of multiplier, are studied. Each type of algorithm is analyzed from the view of parallel and distributed respectively, and algorithms of the same type are compared by their model type, input data characteristic, algorithm evaluation and parallel communication mode. In addition, the implementations and applications of the optimization algorithm on representative scalable machine learning platforms are analyzed. Meanwhile, all the optimization algorithms introduced in this paper are categorized by a hierarchical classification diagram, which can be used as a tool to select the appropriate optimization algorithm according to the objective function type, and also to cross explore how to apply optimization algorithms to the new objective function type. Finally, the problems of the existing optimization algorithms are discussed, and the possible solutions and the future research directions are proposed. © Copyright 2018, Institute of Software, the Chinese Academy of Sciences. All rights reserved.
引用
收藏
页码:109 / 130
相关论文
共 50 条
  • [1] Parallel and Distributed Machine Learning Algorithms for Scalable Big Data Analytics
    Bal, Henri
    Pal, Arindam
    [J]. FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2020, 108 : 1159 - 1161
  • [2] Parallelizing Machine Learning Optimization Algorithms on Distributed Data-Parallel Platforms with Parameter Server
    Gu, Rong
    Fan, Shiqing
    Hu, Qiu
    Yuan, Chunfeng
    Huang, Yihua
    [J]. 2018 IEEE 24TH INTERNATIONAL CONFERENCE ON PARALLEL AND DISTRIBUTED SYSTEMS (ICPADS 2018), 2018, : 126 - 133
  • [3] A Survey of Distributed and Parallel Extreme Learning Machine for Big Data
    Wang, Zhiqiong
    Sui, Ling
    Xin, Junchang
    Qu, Luxuan
    Yao, Yudong
    [J]. IEEE ACCESS, 2020, 8 : 201247 - 201258
  • [4] Scalable and Parallel Machine Learning Algorithms for Statistical Data Mining - Practice & Experience
    Riedel, M.
    Goetz, M.
    Richerzhagen, M.
    Glock, P.
    Bodenstein, C.
    Memon, A. S.
    Memon, M. S.
    [J]. 2015 8TH INTERNATIONAL CONVENTION ON INFORMATION AND COMMUNICATION TECHNOLOGY, ELECTRONICS AND MICROELECTRONICS (MIPRO), 2015, : 204 - 209
  • [5] Communication Optimization Algorithms for Distributed Deep Learning Systems: A Survey
    Yu, Enda
    Dong, Dezun
    Liao, Xiangke
    [J]. IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2023, 34 (12) : 3294 - 3308
  • [6] Dynamic Distributed and Parallel Machine Learning algorithms for big data mining processing
    Djafri, Laouni
    [J]. DATA TECHNOLOGIES AND APPLICATIONS, 2022, 56 (04) : 558 - 601
  • [7] Boosting algorithms for parallel and distributed learning
    Lazarevic, A
    Obradovic, Z
    [J]. DISTRIBUTED AND PARALLEL DATABASES, 2002, 11 (02) : 203 - 229
  • [8] Boosting Algorithms for Parallel and Distributed Learning
    Aleksandar Lazarevic
    Zoran Obradovic
    [J]. Distributed and Parallel Databases, 2002, 11 : 203 - 229
  • [9] A Survey on Distributed Machine Learning
    Verbraeken, Joost
    Wolting, Matthijs
    Katzy, Jonathan
    Kloppenburg, Jeroen
    Verbelen, Tim
    Rellermeyer, Jan S.
    [J]. ACM COMPUTING SURVEYS, 2020, 53 (02)
  • [10] A survey of checkpointing algorithms for parallel and distributed computers
    Kalaiselvi, S
    Rajaraman, V
    [J]. SADHANA-ACADEMY PROCEEDINGS IN ENGINEERING SCIENCES, 2000, 25 (5): : 489 - 510