Communication-efficient distributed cubic Newton with compressed lazy Hessian

被引:0
|
作者
Zhang, Zhen [1 ]
Che, Keqin [1 ]
Yang, Shaofu [1 ]
Xu, Wenying [2 ]
机构
[1] Southeast Univ, Sch Comp Sci & Engn, Nanjing, Jiangsu, Peoples R China
[2] Southeast Univ, Sch Math, Nanjing, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
Distributed optimization; Efficient communication; Second-order algorithms; Cubic Newton method; NEURODYNAMIC APPROACH;
D O I
10.1016/j.neunet.2024.106212
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, second -order distributed optimization algorithms have been becoming a research hot in distributed learning, due to their faster convergence rate than the first -order algorithms. However, second -order algorithms always suffer from serious communication bottleneck. To conquer such challenge, we propose communicationefficient second -order distributed optimization algorithms in the parameter -server framework, by incorporating cubic Newton methods with compressed lazy Hessian. Specifically, our algorithms require each worker communicate compressed Hessians with the server only at some particular iterations, which can save both communication bits and communication rounds. For non -convex problems, we theoretically prove that our algorithms can reduce the communication cost comparing to the state-of-the-art second -order algorithms, while maintaining the same iteration complexity order O ( e -3 / 2 ) as the centralized cubic Newton methods. By further using gradient regularization technique, our algorithms can achieve global convergence for convex problems. Moreover, for strongly convex problems, our algorithms can achieve local superlinear convergence rate without any requirement on initial conditions. Finally, numerical experiments are conducted to show the high efficiency of the proposed algorithms.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Communication-Efficient Distributed SGD With Compressed Sensing
    Tang, Yujie
    Ramanathan, Vikram
    Zhang, Junshan
    Li, Na
    IEEE CONTROL SYSTEMS LETTERS, 2022, 6 : 2054 - 2059
  • [2] Communication-Efficient Distributed Cooperative Learning With Compressed Beliefs
    Toghani, Mohammad Taha
    Uribe, Cesar A.
    IEEE TRANSACTIONS ON CONTROL OF NETWORK SYSTEMS, 2022, 9 (03): : 1215 - 1226
  • [3] Compressed Distributed Gradient Descent: Communication-Efficient Consensus over Networks
    Zhang, Xin
    Liu, Jia
    Zhu, Zhengyuan
    Bentley, Elizabeth S.
    IEEE CONFERENCE ON COMPUTER COMMUNICATIONS (IEEE INFOCOM 2019), 2019, : 2431 - 2439
  • [4] Communication-Efficient Distributed Optimization using an Approximate Newton-type Method
    Shamir, Ohad
    Srebro, Nathan
    Zhang, Tong
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 2), 2014, 32 : 1000 - 1008
  • [5] AC-SGD: Adaptively Compressed SGD for Communication-Efficient Distributed Learning
    Yan, Guangfeng
    Li, Tan
    Huang, Shao-Lun
    Lan, Tian
    Song, Linqi
    IEEE JOURNAL ON SELECTED AREAS IN COMMUNICATIONS, 2022, 40 (09) : 2678 - 2693
  • [6] Communication-Efficient Distributed Eigenspace Estimation
    Charisopoulos, Vasileios
    Benson, Austin R.
    Damle, Anil
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2021, 3 (04): : 1067 - 1092
  • [7] FAST AND COMMUNICATION-EFFICIENT DISTRIBUTED PCA
    Gang, Arpita
    Raja, Haroon
    Bajwa, Waheed U.
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 7450 - 7454
  • [8] Communication-efficient distributed oblivious transfer
    Beimel, Amos
    Chee, Yeow Meng
    Wang, Huaxiong
    Zhang, Liang Feng
    JOURNAL OF COMPUTER AND SYSTEM SCIENCES, 2012, 78 (04) : 1142 - 1157
  • [9] Communication-Efficient Distributed Skyline Computation
    Zhang, Haoyu
    Zhang, Qin
    CIKM'17: PROCEEDINGS OF THE 2017 ACM CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, 2017, : 437 - 446
  • [10] Communication-efficient Distributed SGD with Sketching
    Ivkin, Nikita
    Rothchild, Daniel
    Ullah, Enayat
    Braverman, Vladimir
    Stoica, Ion
    Arora, Raman
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32