Accelerated gradient methods for nonconvex nonlinear and stochastic programming

被引:301
|
作者
Ghadimi, Saeed [1 ]
Lan, Guanghui [1 ]
机构
[1] Univ Florida, Dept Ind & Syst Engn, Gainesville, FL 32611 USA
基金
美国国家科学基金会;
关键词
Nonconvex optimization; Stochastic programming; Accelerated gradient; Complexity; APPROXIMATION ALGORITHMS; COMPOSITE OPTIMIZATION; COMPLEXITY;
D O I
10.1007/s10107-015-0871-8
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
In this paper, we generalize the well-known Nesterov's accelerated gradient (AG) method, originally designed for convex smooth optimization, to solve nonconvex and possibly stochastic optimization problems. We demonstrate that by properly specifying the stepsize policy, the AG method exhibits the best known rate of convergence for solving general nonconvex smooth optimization problems by using first-order information, similarly to the gradient descent method. We then consider an important class of composite optimization problems and show that the AG method can solve them uniformly, i.e., by using the same aggressive stepsize policy as in the convex case, even if the problem turns out to be nonconvex. We demonstrate that the AG method exhibits an optimal rate of convergence if the composite problem is convex, and improves the best known rate of convergence if the problem is nonconvex. Based on the AG method, we also present new nonconvex stochastic approximation methods and show that they can improve a few existing rates of convergence for nonconvex stochastic optimization. To the best of our knowledge, this is the first time that the convergence of the AG method has been established for solving nonconvex nonlinear programming in the literature.
引用
收藏
页码:59 / 99
页数:41
相关论文
共 50 条
  • [1] Accelerated gradient methods for nonconvex nonlinear and stochastic programming
    Saeed Ghadimi
    Guanghui Lan
    [J]. Mathematical Programming, 2016, 156 : 59 - 99
  • [2] Accelerated Proximal Gradient Methods for Nonconvex Programming
    Li, Huan
    Lin, Zhouchen
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [3] Stagewise Accelerated Stochastic Gradient Methods for Nonconvex Optimization
    Jia, Cui
    Cui, Zhuoxu
    [J]. MATHEMATICS, 2024, 12 (11)
  • [4] pbSGD: Powered Stochastic Gradient Descent Methods for Accelerated Nonconvex Optimization
    Zhou, Beitong
    Liu, Jun
    Sun, Weigao
    Chen, Ruijuan
    Tomlin, Claire
    Yuan, Ye
    [J]. PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 3258 - 3266
  • [5] Stochastic perturbation of reduced gradient & GRG methods for nonconvex programming problems
    El Mouatasim, Abdelkrim
    Ellaia, Rachid
    Souza de Cursi, Eduardo
    [J]. APPLIED MATHEMATICS AND COMPUTATION, 2014, 226 : 198 - 211
  • [6] Adaptivity of Stochastic Gradient Methods for Nonconvex Optimization
    Horvath, Samuel
    Lei, Lihua
    Richtarik, Peter
    Jordan, Michael I.
    [J]. SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2022, 4 (02): : 634 - 648
  • [7] Accelerated gradient methods for sparse statistical learning with nonconvex penalties
    Yang, Kai
    Asgharian, Masoud
    Bhatnagar, Sahir
    [J]. STATISTICS AND COMPUTING, 2024, 34 (01)
  • [8] Accelerated gradient methods for sparse statistical learning with nonconvex penalties
    Kai Yang
    Masoud Asgharian
    Sahir Bhatnagar
    [J]. Statistics and Computing, 2024, 34
  • [9] Stochastic proximal gradient methods for nonconvex problems in Hilbert spaces
    Caroline Geiersbach
    Teresa Scarinci
    [J]. Computational Optimization and Applications, 2021, 78 : 705 - 740
  • [10] STOCHASTIC FIRST- AND ZEROTH-ORDER METHODS FOR NONCONVEX STOCHASTIC PROGRAMMING
    Ghadimi, Saeed
    Lan, Guanghui
    [J]. SIAM JOURNAL ON OPTIMIZATION, 2013, 23 (04) : 2341 - 2368