Accelerated gradient sliding for structured convex optimization

被引:0
|
作者
Guanghui Lan
Yuyuan Ouyang
机构
[1] Georgia Institute of Technology,H. Milton Stewart School of Industrial and Systems Engineering
[2] Clemson University,School of Mathematical and Statistical Sciences
关键词
Convex programming; Accelerated gradient sliding; Structure; Complexity; Nesterov’s method; 90C25; 90C06; 49M37;
D O I
暂无
中图分类号
学科分类号
摘要
Our main goal in this paper is to show that one can skip gradient computations for gradient descent type methods applied to certain structured convex programming (CP) problems. To this end, we first present an accelerated gradient sliding (AGS) method for minimizing the summation of two smooth convex functions with different Lipschitz constants. We show that the AGS method can skip the gradient computation for one of these smooth components without slowing down the overall optimal rate of convergence. This result is much sharper than the classic black-box CP complexity results especially when the difference between the two Lipschitz constants associated with these components is large. We then consider an important class of bilinear saddle point problem whose objective function is given by the summation of a smooth component and a nonsmooth one with a bilinear saddle point structure. Using the aforementioned AGS method for smooth composite optimization and Nesterov’s smoothing technique, we show that one only needs O(1/ε)\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${{\mathcal{O}}}(1/\sqrt{\varepsilon })$$\end{document} gradient computations for the smooth component while still preserving the optimal O(1/ε)\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$${{\mathcal{O}}}(1/\varepsilon )$$\end{document} overall iteration complexity for solving these saddle point problems. We demonstrate that even more significant savings on gradient computations can be obtained for strongly convex smooth and bilinear saddle point problems.
引用
收藏
页码:361 / 394
页数:33
相关论文
共 50 条
  • [1] Accelerated gradient sliding for structured convex optimization
    Lan, Guanghui
    Ouyang, Yuyuan
    [J]. COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2022, 82 (02) : 361 - 394
  • [2] CONDITIONAL GRADIENT SLIDING FOR CONVEX OPTIMIZATION
    Lan, Guanghui
    Zhou, Yi
    [J]. SIAM JOURNAL ON OPTIMIZATION, 2016, 26 (02) : 1379 - 1409
  • [3] UNIVERSAL CONDITIONAL GRADIENT SLIDING FOR CONVEX OPTIMIZATION
    Ouyang, Yuyuan
    Squires, Trevor
    [J]. SIAM JOURNAL ON OPTIMIZATION, 2023, 33 (04) : 2962 - 2987
  • [4] Nesterov Accelerated Shuffling Gradient Method for Convex Optimization
    Tran, Trang H.
    Scheinberg, Katya
    Nguyen, Lam M.
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [5] A NOTE ON THE (ACCELERATED) PROXIMAL GRADIENT METHOD FOR COMPOSITE CONVEX OPTIMIZATION
    Li, Qingjing
    Tan, Li
    Guo, Ke
    [J]. JOURNAL OF NONLINEAR AND CONVEX ANALYSIS, 2022, 23 (12) : 2847 - 2857
  • [6] A feasible smoothing accelerated projected gradient method for nonsmooth convex optimization
    Nishioka, Akatsuki
    Kanno, Yoshihiro
    [J]. OPERATIONS RESEARCH LETTERS, 2024, 57
  • [7] An accelerated minimal gradient method with momentum for strictly convex quadratic optimization
    Oviedo, Harry
    Dalmau, Oscar
    Herrera, Rafael
    [J]. BIT NUMERICAL MATHEMATICS, 2022, 62 (02) : 591 - 606
  • [8] Accelerated Bregman proximal gradient methods for relatively smooth convex optimization
    Hanzely, Filip
    Richtarik, Peter
    Xiao, Lin
    [J]. COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2021, 79 (02) : 405 - 440
  • [9] A unified variance-reduced accelerated gradient method for convex optimization
    Lan, Guanghui
    Li, Zhize
    Zhou, Yi
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [10] An accelerated minimal gradient method with momentum for strictly convex quadratic optimization
    Harry Oviedo
    Oscar Dalmau
    Rafael Herrera
    [J]. BIT Numerical Mathematics, 2022, 62 : 591 - 606