Near-optimal Local Convergence of Alternating Gradient Descent-Ascent for Minimax Optimization

被引:0
|
作者
Zhang, Guodong [1 ]
Wang, Yuanhao [2 ]
Lessard, Laurent [3 ]
Grosse, Roger [1 ]
机构
[1] Univ Toronto, Toronto, ON, Canada
[2] Princeton Univ, Princeton, NJ 08544 USA
[3] Northeastern Univ, Boston, MA 02115 USA
关键词
VARIATIONAL INEQUALITY; ALGORITHMS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Smooth minimax games often proceed by simultaneous or alternating gradient updates. Although algorithms with alternating updates are commonly used in practice, the majority of existing theoretical analyses focus on simultaneous algorithms for convenience of analysis. In this paper, we study alternating gradient descent-ascent (Alt-GDA) in minimax games and show that Alt-GDA is superior to its simultaneous counterpart (SimGDA) in many settings. We prove that AltGDA achieves a near-optimal local convergence rate for strongly convex-strongly concave (SCSC) problems while Sim-GDA converges at a much slower rate. To our knowledge, this is the first result of any setting showing that Alt-GDA converges faster than Sim-GDA by more than a constant. We further adapt the theory of integral quadratic constraints (IQC) and show that Alt-GDA attains the same rate globally for a subclass of SCSC minimax problems. Empirically, we demonstrate that alternating updates speed up GAN training significantly and the use of optimism only helps for simultaneous algorithms.
引用
收藏
页数:21
相关论文
共 50 条
  • [41] Distributed Stochastic Gradient Descent: Nonconvexity, Nonsmoothness, and Convergence to Local Minima
    Swenson, Brian
    Murray, Ryan
    Poor, H. Vincent
    Kar, Soummya
    JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [42] PA-GD: On the Convergence of Perturbed Alternating Gradient Descent to Second-Order Stationary Points for Structured Nonconvex Optimization
    Lu, Songtao
    Hong, Mingyi
    Wang, Zhengdao
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [43] A Class of Near-Optimal Local Minima for Witsenhausen's Problem
    Ajorlou, Amir
    Jadbabaie, Ali
    2019 IEEE 58TH CONFERENCE ON DECISION AND CONTROL (CDC), 2019, : 5014 - 5019
  • [44] Visiting near-optimal solutions using local search algorithms
    Jacobson, Sheldon H.
    Hall, Shane N.
    McLay, Laura A.
    COMPSTAT 2006: PROCEEDINGS IN COMPUTATIONAL STATISTICS, 2006, : 471 - +
  • [45] CONVERGENCE RATE OF GRADIENT DESCENT METHOD FOR MULTI-OBJECTIVE OPTIMIZATION
    Zeng, Liaoyuan
    Dai, Yuhong
    Huang, Yakui
    JOURNAL OF COMPUTATIONAL MATHEMATICS, 2019, 37 (05) : 689 - 703
  • [46] The convergence properties of some descent conjugate gradient algorithms for optimization models
    Sulaiman, I. M.
    Mamat, M.
    Owoyemi, A. E.
    Ghazali, P. L.
    Rivaie, M.
    Malik, M.
    JOURNAL OF MATHEMATICS AND COMPUTER SCIENCE-JMCS, 2021, 22 (03): : 204 - 215
  • [47] Nonlinear Optimization Method Based on Stochastic Gradient Descent for Fast Convergence
    Watanabe, Takahiro
    Iima, Hitoshi
    2018 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2018, : 4198 - 4203
  • [48] Explicit Near-Optimal Guidance and Pulse-Modulated Control for Lunar Descent and Touchdown
    M. Pontani
    F. Celani
    S. Carletta
    Aerotecnica Missili & Spazio, 2022, 101 (4): : 361 - 370
  • [49] Complexity of near-optimal robust versions of multilevel optimization problems
    Mathieu Besançon
    Miguel F. Anjos
    Luce Brotcorne
    Optimization Letters, 2021, 15 : 2597 - 2610
  • [50] Near-Optimal Bayesian Ambiguity Sets for Distributionally Robust Optimization
    Gupta, Vishal
    MANAGEMENT SCIENCE, 2019, 65 (09) : 4242 - 4260