Near-optimal Local Convergence of Alternating Gradient Descent-Ascent for Minimax Optimization

被引:0
|
作者
Zhang, Guodong [1 ]
Wang, Yuanhao [2 ]
Lessard, Laurent [3 ]
Grosse, Roger [1 ]
机构
[1] Univ Toronto, Toronto, ON, Canada
[2] Princeton Univ, Princeton, NJ 08544 USA
[3] Northeastern Univ, Boston, MA 02115 USA
关键词
VARIATIONAL INEQUALITY; ALGORITHMS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Smooth minimax games often proceed by simultaneous or alternating gradient updates. Although algorithms with alternating updates are commonly used in practice, the majority of existing theoretical analyses focus on simultaneous algorithms for convenience of analysis. In this paper, we study alternating gradient descent-ascent (Alt-GDA) in minimax games and show that Alt-GDA is superior to its simultaneous counterpart (SimGDA) in many settings. We prove that AltGDA achieves a near-optimal local convergence rate for strongly convex-strongly concave (SCSC) problems while Sim-GDA converges at a much slower rate. To our knowledge, this is the first result of any setting showing that Alt-GDA converges faster than Sim-GDA by more than a constant. We further adapt the theory of integral quadratic constraints (IQC) and show that Alt-GDA attains the same rate globally for a subclass of SCSC minimax problems. Empirically, we demonstrate that alternating updates speed up GAN training significantly and the use of optimism only helps for simultaneous algorithms.
引用
收藏
页数:21
相关论文
共 50 条
  • [1] Local Convergence of Gradient Descent-Ascent for Training Generative Adversarial Networks
    Becker, Evan
    Pandit, Parthe
    Rangan, Sundeep
    Fletcher, Alyson K.
    FIFTY-SEVENTH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, IEEECONF, 2023, : 892 - 896
  • [2] Convergence Rates of Gradient Descent-Ascent Dynamics Under Delays in Solving Nonconvex Min-Max Optimization
    Do, Duy Anh
    Doan, Thinh T.
    2024 European Control Conference, ECC 2024, 2024, : 2748 - 2753
  • [3] Stochastic Smoothed Gradient Descent Ascent for Federated Minimax Optimization
    Shen, Wei
    Huang, Minhui
    Zhang, Jiawei
    Shen, Cong
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [4] On Convergence of Gradient Descent Ascent: A Tight Local Analysis
    Li, Haochuan
    Farnia, Farzan
    Das, Subhro
    Jadbabaie, Ali
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [5] Stochastic Gradient Descent-Ascent and Consensus Optimization for Smooth Games: Convergence Analysis under Expected Co-coercivity
    Loizou, Nicolas
    Berard, Hugo
    Gidel, Gauthier
    Mitliagkas, Ioannis
    Lacoste-Julien, Simon
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [6] AdaGDA: Faster Adaptive Gradient Descent Ascent Methods for Minimax Optimization
    Huang, Feihu
    Wu, Xidong
    Hu, Zhengmian
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 206, 2023, 206
  • [7] Accelerated Proximal Alternating Gradient-Descent-Ascent for Nonconvex Minimax Machine Learning
    Chen, Ziyi
    Ma, Shaocong
    Zhou, Yi
    2022 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY, ISIT, 2022, : 672 - 677
  • [8] Universal Gradient Descent Ascent Method for Nonconvex-Nonconcave Minimax Optimization
    Zheng, Taoli
    Zhu, Linglingzhi
    So, Anthony Man-Cho
    Blanchet, Jose
    Li, Jiajin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [9] Universal Gradient Descent Ascent Method for Nonconvex-Nonconcave Minimax Optimization
    Zheng, Taoli
    Zhu, Linglingzhi
    So, Anthony Man-Cho
    Blanchet, José
    Li, Jiajin
    Advances in Neural Information Processing Systems, 2023, 36 : 54075 - 54110
  • [10] Two-Timescale Gradient Descent Ascent Algorithms for Nonconvex Minimax Optimization
    Lin, Tianyi
    Jin, Chi
    Jordan, Michael I.
    Journal of Machine Learning Research, 2025, 26