Communication Efficient Distributed Newton Method with Fast Convergence Rates

被引:1
|
作者
Liu, Chengchang [1 ]
Chen, Lesi [2 ]
Luo, Luo [2 ]
Lui, John C. S. [1 ]
机构
[1] Chinese Univ Hong Kong, Dept Comp Sci & Engn, Hong Kong, Peoples R China
[2] Fudan Univ, Sch Data Sci, Shanghai, Peoples R China
基金
中国国家自然科学基金;
关键词
Distributed Optimization; Second-Order Methods; CUBIC REGULARIZATION;
D O I
10.1145/3580305.3599280
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We propose a communication and computation efficient second-order method for distributed optimization. For each iteration, our method only requires O(d) communication complexity, where d is the problem dimension. We also provide theoretical analysis to show the proposed method has the similar convergence rate as the classical second-order optimization algorithms. Concretely, our method can find (epsilon, root dL epsilon)-second-order stationary points for nonconvex problem by O(root dL epsilon(-3/2)) iterations, where L is the Lipschitz constant of Hessian. Moreover, it enjoys a local superlinear convergence under the strongly-convex assumption. Experiments on both convex and nonconvex problems show that our proposed method performs significantly better than baselines.
引用
收藏
页码:1406 / 1416
页数:11
相关论文
共 50 条