A STOCHASTIC QUASI-NEWTON METHOD FOR LARGE-SCALE OPTIMIZATION

被引:225
|
作者
Byrd, R. H. [1 ]
Hansen, S. L. [2 ]
Nocedal, Jorge [3 ]
Singer, Y. [4 ]
机构
[1] Univ Colorado, Dept Comp Sci, Boulder, CO 80309 USA
[2] Northwestern Univ, Dept Engn Sci & Appl Math, Evanston, IL 60202 USA
[3] Northwestern Univ, Dept Ind Engn & Management Sci, Evanston, IL 60208 USA
[4] Google Res, Mountain View, CA USA
基金
美国国家科学基金会;
关键词
stochastic optimization; quasi-Newton; sub sampling; large scale optimization; SUBGRADIENT METHODS; GRADIENT;
D O I
10.1137/140954362
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
The question of how to incorporate curvature information into stochastic approximation methods is challenging. The direct application of classical quasi-Newton updating techniques for deterministic optimization leads to noisy curvature estimates that have harmful effects on the robustness of the iteration. In this paper, we propose a stochastic quasi-Newton method that is efficient, robust, and scalable. It employs the classical BFGS update formula in its limited memory form, and is based on the observation that it is beneficial to collect curvature information pointwise, and at spaced intervals. One way to do this is through (subsampled) Hessian-vector products. This technique differs from the classical approach that would compute differences of gradients at every iteration, and where controlling the quality of the curvature estimates can be difficult. We present numerical results on problems arising in machine learning that suggest that the proposed method shows much promise.
引用
收藏
页码:1008 / 1031
页数:24
相关论文
共 50 条
  • [41] A PROJECTIVE QUASI-NEWTON METHOD FOR NONLINEAR OPTIMIZATION
    ZHANG, JZ
    ZHU, DT
    [J]. JOURNAL OF COMPUTATIONAL AND APPLIED MATHEMATICS, 1994, 53 (03) : 291 - 307
  • [42] Regularization of limited memory quasi-Newton methods for large-scale nonconvex minimization
    Kanzow, Christian
    Steck, Daniel
    [J]. MATHEMATICAL PROGRAMMING COMPUTATION, 2023, 15 (03) : 417 - 444
  • [43] Multifidelity Quasi-Newton Method for Design Optimization
    Bryson, Dean E.
    Rumpfkeil, Markus P.
    [J]. AIAA JOURNAL, 2018, 56 (10) : 4074 - 4086
  • [44] Regularization of limited memory quasi-Newton methods for large-scale nonconvex minimization
    Christian Kanzow
    Daniel Steck
    [J]. Mathematical Programming Computation, 2023, 15 : 417 - 444
  • [45] Quasi-Newton's method for multiobjective optimization
    Povalej, Ziga
    [J]. JOURNAL OF COMPUTATIONAL AND APPLIED MATHEMATICS, 2014, 255 : 765 - 777
  • [46] An Improved Quasi-Newton Method for Unconstrained Optimization
    Fei Pusheng
    Chen Zhong (Department of Mathematics
    [J]. Wuhan University Journal of Natural Sciences, 1996, (01) : 35 - 37
  • [47] Stochastic Quasi-Newton Methods
    Mokhtari, Aryan
    Ribeiro, Alejandro
    [J]. PROCEEDINGS OF THE IEEE, 2020, 108 (11) : 1906 - 1922
  • [48] A Variable Sample-size Stochastic Quasi-Newton Method for Smooth and Nonsmooth Stochastic Convex Optimization
    Jalilzadeh, Afrooz
    Nedic, Angelia
    Shanbhag, Uday V.
    Yousefian, Farzad
    [J]. 2018 IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2018, : 4097 - 4102
  • [49] Reduced quasi-Newton method for simultaneous design and optimization
    Adel Hamdi
    Andreas Griewank
    [J]. Computational Optimization and Applications, 2011, 49 : 521 - 548
  • [50] A new regularized quasi-Newton method for unconstrained optimization
    Zhang, Hao
    Ni, Qin
    [J]. OPTIMIZATION LETTERS, 2018, 12 (07) : 1639 - 1658