Stochastic quasi-Newton methods for non-strongly convex problems: convergence and rate analysis

被引:0
|
作者
Yousefian, Farzad [1 ]
Nedic, Angelia [2 ]
Shanbhag, Uday V. [3 ]
机构
[1] Oklahoma State Univ, Sch Ind Engn & Management, Stillwater, OK 74078 USA
[2] Univ Illinois, Ind & Enterprise Syst Engn, Urbana, IL 61801 USA
[3] Penn State Univ, Ind & Mfg Engn, University Pk, PA 16802 USA
关键词
BFGS METHOD; APPROXIMATION;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Motivated by applications in optimization and machine learning, we consider stochastic quasi-Newton (SQN) methods for solving stochastic optimization problems. In the literature, the convergence analysis of these algorithms relies on strong convexity of the objective function. To our knowledge, no rate statements exist in the absence of this assumption. Motivated by this gap, we allow the objective function to be merely convex and develop a regularized SQN method. In this scheme, both the gradient mapping and the Hessian approximation are regularized at each iteration and updated alternatively. Unlike the classical regularization schemes, we allow the regularization parameter to be updated iteratively and decays to zero. Under suitable assumptions on the stepsize and regularization parameters, we show that the function value converges to its optimal value in both an almost sure and an expected-value sense. In each case, a set of regularization and steplength sequences is provided under which convergence may be guaranteed. Moreover, the rate of convergence is derived in terms of function value. Our empirical analysis on a binary classification problem shows that the proposed scheme performs well compared to both classical regularized SQN and stochastic approximation schemes.
引用
收藏
页码:4496 / 4503
页数:8
相关论文
共 50 条
  • [1] On the Convergence Rate of Quasi-Newton Methods on Strongly Convex Functions with Lipschitz Gradient
    Krutikov, Vladimir
    Tovbis, Elena
    Stanimirovic, Predrag
    Kazakovtsev, Lev
    [J]. MATHEMATICS, 2023, 11 (23)
  • [2] ON STOCHASTIC AND DETERMINISTIC QUASI-NEWTON METHODS FOR NONSTRONGLY CONVEX OPTIMIZATION: ASYMPTOTIC CONVERGENCE AND RATE ANALYSIS
    Yousefian, Farzad
    Nedic, Angelia
    Shanbhag, Uday V.
    [J]. SIAM JOURNAL ON OPTIMIZATION, 2020, 30 (02) : 1144 - 1172
  • [3] GLOBAL CONVERGENCE OF A CLASS OF QUASI-NEWTON METHODS ON CONVEX PROBLEMS
    BYRD, RH
    NOCEDAL, J
    YUAN, YX
    [J]. SIAM JOURNAL ON NUMERICAL ANALYSIS, 1987, 24 (05) : 1171 - 1190
  • [4] On the convergence of quasi-Newton methods for nonsmooth problems
    Lopes, VLR
    Martinez, JM
    [J]. NUMERICAL FUNCTIONAL ANALYSIS AND OPTIMIZATION, 1995, 16 (9-10) : 1193 - 1209
  • [5] Stochastic proximal quasi-Newton methods for non-convex composite optimization
    Wang, Xiaoyu
    Wang, Xiao
    Yuan, Ya-xiang
    [J]. OPTIMIZATION METHODS & SOFTWARE, 2019, 34 (05): : 922 - 948
  • [6] Stochastic Quasi-Newton Methods
    Mokhtari, Aryan
    Ribeiro, Alejandro
    [J]. PROCEEDINGS OF THE IEEE, 2020, 108 (11) : 1906 - 1922
  • [7] Linear convergence of first order methods for non-strongly convex optimization
    Necoara, I.
    Nesterov, Yu.
    Glineur, F.
    [J]. MATHEMATICAL PROGRAMMING, 2019, 175 (1-2) : 69 - 107
  • [8] Linear convergence of first order methods for non-strongly convex optimization
    I. Necoara
    Yu. Nesterov
    F. Glineur
    [J]. Mathematical Programming, 2019, 175 : 69 - 107
  • [10] On the convergence rate of a quasi-Newton method for inverse eigenvalue problems
    Chan, RH
    Xu, SF
    Zhou, HM
    [J]. SIAM JOURNAL ON NUMERICAL ANALYSIS, 1999, 36 (02) : 436 - 441