Stochastic Recursive Gradient Descent Ascent for Stochastic Nonconvex-Strongly-Concave Minimax Problems

被引:0
|
作者
Luo, Luo [1 ]
Ye, Haishan [2 ]
Huang, Zhichao [1 ]
Zhang, Tong [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Dept Math, Hong Kong, Peoples R China
[2] Chinese Univ Hong Kong, Shenzhen Res Inst Big Data, Shenzhen, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider nonconvex-concave minimax optimization problems of the form min(x) max(y is an element of Y) f (x; y), where f is strongly-concave in y but possibly nonconvex in x and Y is a convex and compact set. We focus on the stochastic setting, where we can only access an unbiased stochastic gradient estimate of f at each iteration. This formulation includes many machine learning applications as special cases such as robust optimization and adversary training. We are interested in finding an O(epsilon)-stationary point of the function Phi(center dot) = max(y is an element of Y) f (center dot, y). The most popular algorithm to solve this problem is stochastic gradient decent ascent, which requires O(kappa 3 epsilon(-4)) stochastic gradient evaluations, where kappa is the condition number. In this paper, we propose a novel method called Stochastic Recursive gradiEnt Descent Ascent (SREDA), which estimates gradients more efficiently using variance reduction. This method achieves the best known stochastic gradient complexity of O(kappa 3 epsilon(-4)), and its dependency on epsilon is optimal for this problem.
引用
收藏
页数:12
相关论文
共 50 条
  • [31] Complexity Lower Bounds for Nonconvex-Strongly-Concave Min-Max Optimization
    Li, Haochuan
    Tian, Yi
    Zhang, Jingzhao
    Jadbabaie, Ali
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [32] High Probability Guarantees for Nonconvex Stochastic Gradient Descent with Heavy Tails
    Li, Shaojie
    Liu, Yong
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [33] Second-Order Guarantees of Stochastic Gradient Descent in Nonconvex Optimization
    Vlaski, Stefan
    Sayed, Ali H.
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2022, 67 (12) : 6489 - 6504
  • [34] pbSGD: Powered Stochastic Gradient Descent Methods for Accelerated Nonconvex Optimization
    Zhou, Beitong
    Liu, Jun
    Sun, Weigao
    Chen, Ruijuan
    Tomlin, Claire
    Yuan, Ye
    PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 3258 - 3266
  • [35] ZEROTH-ORDER STOCHASTIC PROJECTED GRADIENT DESCENT FOR NONCONVEX OPTIMIZATION
    Liu, Sijia
    Li, Xingguo
    Chen, Pin-Yu
    Haupt, Jarvis
    Amini, Lisa
    2018 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP 2018), 2018, : 1179 - 1183
  • [36] Stability and Generalization of Stochastic Gradient Methods for Minimax Problems
    Lei, Yunwen
    Yang, Zhenhuan
    Yang, Tianbao
    Ying, Yiming
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [37] An Optimal Multistage Stochastic Gradient Method for Minimax Problems
    Fallah, Alireza
    Ozdaglar, Asuman
    Pattathil, Sarath
    2020 59TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2020, : 3573 - 3579
  • [38] Zeroth-order algorithms for nonconvex–strongly-concave minimax problems with improved complexities
    Zhongruo Wang
    Krishnakumar Balasubramanian
    Shiqian Ma
    Meisam Razaviyayn
    Journal of Global Optimization, 2023, 87 : 709 - 740
  • [39] STOCHASTIC GRADIENT DESCENT ALGORITHM FOR STOCHASTIC OPTIMIZATION IN SOLVING ANALYTIC CONTINUATION PROBLEMS
    Bao, Feng
    Maier, Thomas
    FOUNDATIONS OF DATA SCIENCE, 2020, 2 (01): : 1 - 17
  • [40] Stochastic Gradient Descent-Ascent: Unified Theory and New Efficient Methods
    Beznosikov, Aleksandr
    Gorbunov, Eduard
    Berard, Hugo
    Loizou, Nicolas
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 206, 2023, 206 : 172 - 235