Accelerated gradient methods for nonconvex nonlinear and stochastic programming

被引:300
|
作者
Ghadimi, Saeed [1 ]
Lan, Guanghui [1 ]
机构
[1] Univ Florida, Dept Ind & Syst Engn, Gainesville, FL 32611 USA
基金
美国国家科学基金会;
关键词
Nonconvex optimization; Stochastic programming; Accelerated gradient; Complexity; APPROXIMATION ALGORITHMS; COMPOSITE OPTIMIZATION; COMPLEXITY;
D O I
10.1007/s10107-015-0871-8
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
In this paper, we generalize the well-known Nesterov's accelerated gradient (AG) method, originally designed for convex smooth optimization, to solve nonconvex and possibly stochastic optimization problems. We demonstrate that by properly specifying the stepsize policy, the AG method exhibits the best known rate of convergence for solving general nonconvex smooth optimization problems by using first-order information, similarly to the gradient descent method. We then consider an important class of composite optimization problems and show that the AG method can solve them uniformly, i.e., by using the same aggressive stepsize policy as in the convex case, even if the problem turns out to be nonconvex. We demonstrate that the AG method exhibits an optimal rate of convergence if the composite problem is convex, and improves the best known rate of convergence if the problem is nonconvex. Based on the AG method, we also present new nonconvex stochastic approximation methods and show that they can improve a few existing rates of convergence for nonconvex stochastic optimization. To the best of our knowledge, this is the first time that the convergence of the AG method has been established for solving nonconvex nonlinear programming in the literature.
引用
收藏
页码:59 / 99
页数:41
相关论文
共 50 条
  • [41] Interior-point methods for nonconvex nonlinear programming: orderings and higher-order methods
    David F. Shanno
    Robert J. Vanderbei
    [J]. Mathematical Programming, 2000, 87 : 303 - 316
  • [42] On the diffusion approximation of nonconvex stochastic gradient descent
    Hu, Wenqing
    Li, Chris Junchi
    Li, Lei
    Liu, Jian-Guo
    [J]. ANNALS OF MATHEMATICAL SCIENCES AND APPLICATIONS, 2019, 4 (01) : 3 - 32
  • [43] Stochastic generalized gradient method for nonconvex nonsmooth stochastic optimization
    Yu. M. Ermol'ev
    V. I. Norkin
    [J]. Cybernetics and Systems Analysis, 1998, 34 : 196 - 215
  • [44] VARIABLE METRIC PROXIMAL STOCHASTIC VARIANCE REDUCED GRADIENT METHODS FOR NONCONVEX NONSMOOTH OPTIMIZATION
    Yu, Tengteng
    Liu, Xin-wei
    Dai, Yu-hong
    Sun, J. I. E.
    [J]. JOURNAL OF INDUSTRIAL AND MANAGEMENT OPTIMIZATION, 2022, 18 (04) : 2611 - 2631
  • [45] Asynchronous Parallel Stochastic Gradient for Nonconvex Optimization
    Lian, Xiangru
    Huang, Yijun
    Li, Yuncheng
    Liu, Ji
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [46] Stochastic generalized gradient method for nonconvex nonsmooth stochastic optimization
    Ermol'ev, YM
    Norkin, VI
    [J]. CYBERNETICS AND SYSTEMS ANALYSIS, 1998, 34 (02) : 196 - 215
  • [47] Accelerated Stochastic Gradient-free and Projection-free Methods
    Huang, Feihu
    Tao, Lue
    Chen, Songcan
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [48] A Lyapunov analysis for accelerated gradient methods: from deterministic to stochastic case
    Laborde, M.
    Oberman, A.
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108 : 602 - 611
  • [49] A note on the accelerated proximal gradient method for nonconvex optimization
    Wang, Huijuan
    Xu, Hong-Kun
    [J]. CARPATHIAN JOURNAL OF MATHEMATICS, 2018, 34 (03) : 449 - 457
  • [50] Planar methods and grossone for the Conjugate Gradient breakdown in nonlinear programming
    Renato De Leone
    Giovanni Fasano
    Yaroslav D. Sergeyev
    [J]. Computational Optimization and Applications, 2018, 71 : 73 - 93