Communication-efficient distributed cubic Newton with compressed lazy Hessian

被引:0
|
作者
Zhang, Zhen [1 ]
Che, Keqin [1 ]
Yang, Shaofu [1 ]
Xu, Wenying [2 ]
机构
[1] Southeast Univ, Sch Comp Sci & Engn, Nanjing, Jiangsu, Peoples R China
[2] Southeast Univ, Sch Math, Nanjing, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
Distributed optimization; Efficient communication; Second-order algorithms; Cubic Newton method; NEURODYNAMIC APPROACH;
D O I
10.1016/j.neunet.2024.106212
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently, second -order distributed optimization algorithms have been becoming a research hot in distributed learning, due to their faster convergence rate than the first -order algorithms. However, second -order algorithms always suffer from serious communication bottleneck. To conquer such challenge, we propose communicationefficient second -order distributed optimization algorithms in the parameter -server framework, by incorporating cubic Newton methods with compressed lazy Hessian. Specifically, our algorithms require each worker communicate compressed Hessians with the server only at some particular iterations, which can save both communication bits and communication rounds. For non -convex problems, we theoretically prove that our algorithms can reduce the communication cost comparing to the state-of-the-art second -order algorithms, while maintaining the same iteration complexity order O ( e -3 / 2 ) as the centralized cubic Newton methods. By further using gradient regularization technique, our algorithms can achieve global convergence for convex problems. Moreover, for strongly convex problems, our algorithms can achieve local superlinear convergence rate without any requirement on initial conditions. Finally, numerical experiments are conducted to show the high efficiency of the proposed algorithms.
引用
收藏
页数:16
相关论文
共 50 条
  • [31] Harvesting Curvatures for Communication-Efficient Distributed Optimization
    Cardoso, Diogo
    Li, Boyue
    Chi, Yuejie
    Xavier, Joao
    2022 56TH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS, 2022, : 749 - 753
  • [32] Communication-efficient distributed mining of association rules
    Schuster, A
    Wolff, R
    DATA MINING AND KNOWLEDGE DISCOVERY, 2004, 8 (02) : 171 - 196
  • [33] Communication-Efficient Computation on Distributed Noisy Datasets
    Zhang, Qin
    SPAA'15: PROCEEDINGS OF THE 27TH ACM SYMPOSIUM ON PARALLELISM IN ALGORITHMS AND ARCHITECTURES, 2015, : 313 - 322
  • [34] CREDO : A Communication-Efficient Distributed Estimation Algorithm
    Sahu, Anit Kumar
    Jakovetic, Dusan
    Kar, Soummya
    2018 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2018, : 516 - 520
  • [35] Communication-efficient estimation for distributed subset selection
    Yan Chen
    Ruipeng Dong
    Canhong Wen
    Statistics and Computing, 2023, 33
  • [36] More communication-efficient distributed sparse learning
    Zhou, Xingcai
    Yang, Guang
    Information Sciences, 2024, 668
  • [37] Communication-efficient distributed covariance sketch, with application to distributed PCA
    Huang, Zengfeng
    Lin, Xuemin
    Zhang, Wenjie
    Zhang, Ying
    Journal of Machine Learning Research, 2021, 22
  • [38] Communication-Efficient Distributed Covariance Sketch, with Application to Distributed PCA
    Huang, Zengfeng
    Lin, Xuemin
    Zhang, Wenjie
    Zhang, Ying
    JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22
  • [39] Communication-Efficient Federated Learning via Quantized Compressed Sensing
    Oh, Yongjeong
    Lee, Namyoon
    Jeon, Yo-Seb
    Poor, H. Vincent
    IEEE TRANSACTIONS ON WIRELESS COMMUNICATIONS, 2023, 22 (02) : 1087 - 1100
  • [40] Communication-Efficient and Resilient Distributed Q-Learning
    Xie, Yijing
    Mou, Shaoshuai
    Sundaram, Shreyas
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (03) : 3351 - 3364