A STOCHASTIC QUASI-NEWTON METHOD FOR LARGE-SCALE OPTIMIZATION

被引:225
|
作者
Byrd, R. H. [1 ]
Hansen, S. L. [2 ]
Nocedal, Jorge [3 ]
Singer, Y. [4 ]
机构
[1] Univ Colorado, Dept Comp Sci, Boulder, CO 80309 USA
[2] Northwestern Univ, Dept Engn Sci & Appl Math, Evanston, IL 60202 USA
[3] Northwestern Univ, Dept Ind Engn & Management Sci, Evanston, IL 60208 USA
[4] Google Res, Mountain View, CA USA
基金
美国国家科学基金会;
关键词
stochastic optimization; quasi-Newton; sub sampling; large scale optimization; SUBGRADIENT METHODS; GRADIENT;
D O I
10.1137/140954362
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
The question of how to incorporate curvature information into stochastic approximation methods is challenging. The direct application of classical quasi-Newton updating techniques for deterministic optimization leads to noisy curvature estimates that have harmful effects on the robustness of the iteration. In this paper, we propose a stochastic quasi-Newton method that is efficient, robust, and scalable. It employs the classical BFGS update formula in its limited memory form, and is based on the observation that it is beneficial to collect curvature information pointwise, and at spaced intervals. One way to do this is through (subsampled) Hessian-vector products. This technique differs from the classical approach that would compute differences of gradients at every iteration, and where controlling the quality of the curvature estimates can be difficult. We present numerical results on problems arising in machine learning that suggest that the proposed method shows much promise.
引用
收藏
页码:1008 / 1031
页数:24
相关论文
共 50 条
  • [1] A Stochastic Quasi-Newton Method for Large-Scale Nonconvex Optimization With Applications
    Chen, Huiming
    Wu, Ho-Chun
    Chan, Shing-Chow
    Lam, Wong-Hing
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (11) : 4776 - 4790
  • [2] Fast large-scale optimization by unifying stochastic gradient and quasi-Newton methods
    Sohl-Dickstein, Jascha
    Poole, Ben
    Ganguli, Surya
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 2), 2014, 32 : 604 - 612
  • [3] A Sequential Subspace Quasi-Newton Method for Large-Scale Convex Optimization
    Senov, Aleksandr
    Granichin, Oleg
    Granichina, Olga
    [J]. 2020 AMERICAN CONTROL CONFERENCE (ACC), 2020, : 3627 - 3632
  • [4] Preconditioned Subspace Quasi-Newton Method for Large Scale Optimization
    Sim, Hong Seng
    Leong, Wah June
    Abu Hassan, Malik
    Ismail, Fudziah
    [J]. PERTANIKA JOURNAL OF SCIENCE AND TECHNOLOGY, 2014, 22 (01): : 175 - 192
  • [5] An Overview of Stochastic Quasi-Newton Methods for Large-Scale Machine Learning
    Guo, Tian-De
    Liu, Yan
    Han, Cong-Ying
    [J]. JOURNAL OF THE OPERATIONS RESEARCH SOCIETY OF CHINA, 2023, 11 (02) : 245 - 275
  • [6] An Overview of Stochastic Quasi-Newton Methods for Large-Scale Machine Learning
    Tian-De Guo
    Yan Liu
    Cong-Ying Han
    [J]. Journal of the Operations Research Society of China, 2023, 11 : 245 - 275
  • [7] STOCHASTIC QUASI-NEWTON METHOD FOR NONCONVEX STOCHASTIC OPTIMIZATION
    Wang, Xiao
    Ma, Shiqian
    Goldfarb, Donald
    Liu, Wei
    [J]. SIAM JOURNAL ON OPTIMIZATION, 2017, 27 (02) : 927 - 956
  • [8] Scaling on Diagonal Quasi-Newton Update for Large-Scale Unconstrained Optimization
    Leong, Wah June
    Farid, Mahboubeh
    Abu Hassan, Malik
    [J]. BULLETIN OF THE MALAYSIAN MATHEMATICAL SCIENCES SOCIETY, 2012, 35 (02) : 247 - 256
  • [9] Preconditioning On Subspace Quasi-Newton Method For Large Scale Unconstrained Optimization
    Sim, Hong Seng
    Leong, Wah June
    Ismail, Fudziah
    [J]. STATISTICS AND OPERATIONAL RESEARCH INTERNATIONAL CONFERENCE (SORIC 2013), 2014, 1613 : 297 - 305
  • [10] An efficient augmented memoryless quasi-Newton method for solving large-scale unconstrained optimization problems
    Cheng, Yulin
    Gao, Jing
    [J]. AIMS MATHEMATICS, 2024, 9 (09): : 25232 - 25252