Stochastic Optimization for Nonconvex Problem With Inexact Hessian Matrix, Gradient, and Function

被引:1
|
作者
Liu, Liu [1 ]
Liu, Xuanqing [2 ]
Hsieh, Cho-Jui [3 ]
Tao, Dacheng [4 ]
机构
[1] Beihang Univ, State Key Lab Software Dev Environm, Beijing 100191, Peoples R China
[2] Amazon Web Serv AWS, Seattle, WA 98108 USA
[3] Univ Calif Los Angeles, Dept Comp Sci, Los Angeles, CA 90095 USA
[4] Univ Sydney, Fac Engn, Sydney AI Ctr, Sch Comp Sci, Sydney, NSW 2008, Australia
关键词
Adaptive regularization; stochastic optimization; trust region (TR); ALGORITHMS; REGULARIZATION; COMPLEXITY;
D O I
10.1109/TNNLS.2023.3326177
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Trust region (TR) and adaptive regularization using cubics (ARC) have proven to have some very appealing theoretical properties for nonconvex optimization by concurrently computing function value, gradient, and Hessian matrix to obtain the next search direction and the adjusted parameters. Although stochastic approximations help largely reduce the computational cost, it is challenging to theoretically guarantee the convergence rate. In this article, we explore a family of stochastic TR (STR) and stochastic ARC (SARC) methods that can simultaneously provide inexact computations of the Hessian matrix, gradient, and function values. Our algorithms require much fewer propagations overhead per iteration than TR and ARC. We prove that the iteration complexity to achieve epsilon-approximate second-order optimality is of the same order as the exact computations demonstrated in previous studies. In addition, the mild conditions on inexactness can be met by leveraging a random sampling technology in the finite-sum minimization problem. Numerical experiments with a nonconvex problem support these findings and demonstrate that, with the same or a similar number of iterations, our algorithms require less computational overhead per iteration than current second order methods.
引用
收藏
页码:1 / 13
页数:13
相关论文
共 50 条
  • [1] Inexact Reduced Gradient Methods in Nonconvex Optimization
    Khanh, Pham Duy
    Mordukhovich, Boris S.
    Tran, Dat Ba
    JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 2023, 203 (3) : 2138 - 2178
  • [2] Inexact reduced gradient methods in nonconvex optimization
    Khanh, Pham Duy
    Mordukhovich, Boris S.
    Tran, Dat Ba
    arXiv, 2022,
  • [3] Asynchronous Parallel Stochastic Gradient for Nonconvex Optimization
    Lian, Xiangru
    Huang, Yijun
    Li, Yuncheng
    Liu, Ji
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [4] Adaptivity of Stochastic Gradient Methods for Nonconvex Optimization
    Horvath, Samuel
    Lei, Lihua
    Richtarik, Peter
    Jordan, Michael I.
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2022, 4 (02): : 634 - 648
  • [5] Accelerated inexact composite gradient methods for nonconvex spectral optimization problems
    Weiwei Kong
    Renato D. C. Monteiro
    Computational Optimization and Applications, 2022, 82 : 673 - 715
  • [6] Accelerated inexact composite gradient methods for nonconvex spectral optimization problems
    Kong, Weiwei
    Monteiro, Renato D. C.
    COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2022, 82 (03) : 673 - 715
  • [7] Stochastic inexact augmented Lagrangian method for nonconvex expectation constrained optimization
    Li, Zichong
    Chen, Pin-Yu
    Liu, Sijia
    Lu, Songtao
    Xu, Yangyang
    COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2024, 87 (01) : 117 - 147
  • [8] Stochastic generalized gradient method for nonconvex nonsmooth stochastic optimization
    Yu. M. Ermol'ev
    V. I. Norkin
    Cybernetics and Systems Analysis, 1998, 34 : 196 - 215
  • [9] Stochastic generalized gradient method for nonconvex nonsmooth stochastic optimization
    Ermol'ev, YM
    Norkin, VI
    CYBERNETICS AND SYSTEMS ANALYSIS, 1998, 34 (02) : 196 - 215
  • [10] An inexact proximal gradient algorithm with extrapolation for a class of nonconvex nonsmooth optimization problems
    Jia, Zehui
    Wu, Zhongming
    Dong, Xiaomei
    JOURNAL OF INEQUALITIES AND APPLICATIONS, 2019, 2019 (1)