SKETCHED NEWTON-RAPHSON

被引:0
|
作者
Yuan, Rui [1 ,2 ]
Lazaric, Alessandro [1 ]
GOWER, ROBERT M. [3 ]
机构
[1] Meta AI, LTCI, Télécom Paris, Paris, France
[2] Institut Polytechnique de Paris, Paris, France
[3] CCM, Flatiron Institute, New York,NY,10010, United States
基金
欧洲研究理事会;
关键词
Gradient methods - Nonlinear equations - Optimization - Stochastic models - Stochastic systems;
D O I
10.1137/19M130604X
中图分类号
学科分类号
摘要
We propose a new globally convergent stochastic second-order method. Our starting point is the development of a new sketched Newton-Raphson (SNR) method for solving large scale nonlinear equations of the form F(x) = 0 with F : Rp → Rm. We then show how to design several stochastic second-order optimization methods by rewriting the optimization problem of interest as a system of nonlinear equations and applying SNR. For instance, by applying SNR to find a stationary point of a generalized linear model, we derive completely new and scalable stochastic second-order methods. We show that the resulting method is very competitive as compared to state-of-the-art variance reduced methods. Furthermore, using a variable splitting trick, we also show that the stochastic Newton method (SNM) is a special case of SNR and use this connection to establish the first global convergence theory of SNM. We establish the global convergence of SNR by showing that it is a variant of the online stochastic gradient descent (SGD) method, and then leveraging proof techniques of SGD. As a special case, our theory also provides a new global convergence theory for the original Newton-Raphson method under strictly weaker assumptions as compared to the classic monotone convergence theory. © 2022 Society for Industrial and Applied Mathematics.
引用
收藏
页码:1555 / 1583
相关论文
共 50 条