Faster Stochastic Variance Reduction Methods for Compositional MiniMax Optimization

被引:0
|
作者
Liu, Jin [1 ]
Pan, Xiaokang [1 ]
Duan, Junwen [1 ]
Li, Hong-Dong [1 ]
Li, Youqi [2 ]
Qu, Zhe [1 ]
机构
[1] Cent South Univ, Sch Comp Sci & Engn, Changsha, Peoples R China
[2] Beijing Inst Technol, Sch Comp Sci & Technol, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper delves into the realm of stochastic optimization for compositional minimax optimization-a pivotal challenge across various machine learning domains, including deep AUC and reinforcement learning policy evaluation. Despite its significance, the problem of compositional minimax optimization is still under-explored. Adding to the complexity, current methods of compositional minimax optimization are plagued by sub-optimal complexities or heavy reliance on sizable batch sizes. To respond to these constraints, this paper introduces a novel method, called Nested STOchastic Recursive Momentum (NSTORM), which can achieve the optimal sample complexity and obtain the nearly accuracy solution, matching the existing minimax methods. We also demonstrate that NSTORM can achieve the same sample complexity under the Polyak-Lojasiewicz (PL)-condition-an insightful extension of its capabilities. Yet, NSTORM encounters an issue with its requirement for low learning rates, potentially constraining its real-world applicability in machine learning. To overcome this hurdle, we present ADAptive NSTORM (ADA-NSTORM) with adaptive learning rates. We demonstrate that ADA-NSTORM can achieve the same sample complexity but the experimental results show its more effectiveness. All the proposed complexities indicate that our proposed methods can match lower bounds to existing minimax optimizations, without requiring a large batch size in each iteration. Extensive experiments support the efficiency of our proposed methods.
引用
收藏
页码:13927 / 13935
页数:9
相关论文
共 50 条
  • [1] SpiderBoost and Momentum: Faster Stochastic Variance Reduction Algorithms
    Wang, Zhe
    Ji, Kaiyi
    Zhou, Yi
    Liang, Yingbin
    Tarokh, Vahid
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [2] Stochastic Variance Reduction for Nonconvex Optimization
    Reddi, Sashank J.
    Hefny, Ahmed
    Sra, Suvrit
    Poczos, Barnabas
    Smola, Alex
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [3] Variance Reduction for Faster Non-Convex Optimization
    Allen-Zhu, Zeyuan
    Hazan, Elad
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [4] Stochastic Nested Variance Reduction for Nonconvex Optimization
    Zhou, Dongruo
    Xu, Pan
    Gu, Quanquan
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [5] Stochastic Nested Variance Reduction for Nonconvex Optimization
    Zhou, Dongruo
    Xu, Pan
    Gu, Quanquan
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [6] Stochastic Variance Reduction Methods for Policy Evaluation
    Du, Simon S.
    Chen, Jianshu
    Li, Lihong
    Xiao, Lin
    Zhou, Dengyong
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [7] Faster Stochastic Algorithms for Minimax Optimization under Polyak-Lojasiewicz Conditions
    Chen, Lesi
    Yao, Boyuan
    Luo, Luo
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [8] Faster Stochastic Algorithms for Minimax Optimization under Polyak-Lojasiewicz Conditions
    Chen, Lesi
    Yao, Boyuan
    Luo, Luo
    [J]. Advances in Neural Information Processing Systems, 2022, 35
  • [9] Fast Training Methods for Stochastic Compositional Optimization Problems
    Gao, Hongchang
    Huang, Heng
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [10] Decentralized Stochastic Optimization With Pairwise Constraints and Variance Reduction
    Han, Fei
    Cao, Xuanyu
    Gong, Yi
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2024, 72 : 1960 - 1973