Stochastic Recursive Gradient Descent Ascent for Stochastic Nonconvex-Strongly-Concave Minimax Problems

被引:0
|
作者
Luo, Luo [1 ]
Ye, Haishan [2 ]
Huang, Zhichao [1 ]
Zhang, Tong [1 ]
机构
[1] Hong Kong Univ Sci & Technol, Dept Math, Hong Kong, Peoples R China
[2] Chinese Univ Hong Kong, Shenzhen Res Inst Big Data, Shenzhen, Peoples R China
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider nonconvex-concave minimax optimization problems of the form min(x) max(y is an element of Y) f (x; y), where f is strongly-concave in y but possibly nonconvex in x and Y is a convex and compact set. We focus on the stochastic setting, where we can only access an unbiased stochastic gradient estimate of f at each iteration. This formulation includes many machine learning applications as special cases such as robust optimization and adversary training. We are interested in finding an O(epsilon)-stationary point of the function Phi(center dot) = max(y is an element of Y) f (center dot, y). The most popular algorithm to solve this problem is stochastic gradient decent ascent, which requires O(kappa 3 epsilon(-4)) stochastic gradient evaluations, where kappa is the condition number. In this paper, we propose a novel method called Stochastic Recursive gradiEnt Descent Ascent (SREDA), which estimates gradients more efficiently using variance reduction. This method achieves the best known stochastic gradient complexity of O(kappa 3 epsilon(-4)), and its dependency on epsilon is optimal for this problem.
引用
收藏
页数:12
相关论文
共 50 条
  • [41] SSRGD: Simple Stochastic Recursive Gradient Descent for Escaping Saddle Points
    Li, Zhize
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [42] A Communication-Efficient Stochastic Gradient Descent Algorithm for Distributed Nonconvex Optimization
    Xie, Antai
    Yi, Xinlei
    Wang, Xiaofan
    Cao, Ming
    Ren, Xiaoqiang
    2024 IEEE 18TH INTERNATIONAL CONFERENCE ON CONTROL & AUTOMATION, ICCA 2024, 2024, : 609 - 614
  • [43] Proximal stochastic recursive momentum algorithm for nonsmooth nonconvex optimization problems
    Wang, Zhaoxin
    Wen, Bo
    OPTIMIZATION, 2024, 73 (02) : 481 - 495
  • [44] Stochastic proximal gradient methods for nonconvex problems in Hilbert spaces
    Caroline Geiersbach
    Teresa Scarinci
    Computational Optimization and Applications, 2021, 78 : 705 - 740
  • [45] Stochastic proximal gradient methods for nonconvex problems in Hilbert spaces
    Geiersbach, Caroline
    Scarinci, Teresa
    COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2021, 78 (03) : 705 - 740
  • [46] Alternating Gradient Descent Ascent for Nonconvex Min-Max Problems in Robust Learning and GANs
    Lu, Songtao
    Singh, Rahul
    Chen, Xiangyi
    Chen, Yongxin
    Hong, Mingyi
    CONFERENCE RECORD OF THE 2019 FIFTY-THIRD ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, 2019, : 680 - 684
  • [47] Efficient Smooth Non-Convex Stochastic Compositional Optimization via Stochastic Recursive Gradient Descent
    Hu, Wenqing
    Li, Chris Junchi
    Lian, Xiangru
    Liu, Ji
    Yuan, Huizhuo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [48] Unforgeability in Stochastic Gradient Descent
    Baluta, Teodora
    Nikolic, Ivica
    Jain, Racchit
    Aggarwal, Divesh
    Saxena, Prateek
    PROCEEDINGS OF THE 2023 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, CCS 2023, 2023, : 1138 - 1152
  • [49] Preconditioned Stochastic Gradient Descent
    Li, Xi-Lin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (05) : 1454 - 1466
  • [50] Stochastic gradient descent tricks
    Bottou, Léon
    Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2012, 7700 LECTURE NO : 421 - 436