Variance reduction on general adaptive stochastic mirror descent

被引:1
|
作者
Li, Wenjie [1 ]
Wang, Zhanyu [1 ]
Zhang, Yichen [2 ]
Cheng, Guang [3 ]
机构
[1] Purdue Univ, Dept Stat, W Lafayette, IN 47907 USA
[2] Purdue Univ, Krannert Sch Management, W Lafayette, IN 47907 USA
[3] Univ Calif Los Angeles, Dept Stat, Los Angeles, CA USA
关键词
Variance reduction; Adaptive mirror descent; Nonconvex nonsmooth optimization; General framework; Convergence analysis;
D O I
10.1007/s10994-022-06227-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this work, we propose a simple and generalized algorithmic framework for applying variance reduction to adaptive mirror descent algorithms for faster convergence. We introduce the SVRAMD algorithm, and provide its general convergence analysis in both the nonsmooth nonconvex optimization problem and the generalized P-L conditioned nonconvex optimization problem. We prove that variance reduction can reduce the gradient complexity of all adaptive mirror descent algorithms that satisfy a mild assumption and thus accelerate their convergence. In particular, our general theory implies that variance reduction can be applied to different algorithms with their distinct choices of the proximal function, such as gradient descent with time-varying step sizes, mirror descent with L-1 mirror maps, and self-adaptive algorithms such as AdaGrad and RMSProp. Moreover, the proved convergence rates of SVRAMD recover the existing rates without complicated algorithmic components, which indicates their optimality. Extensive experiments validate our theoretical findings.
引用
收藏
页码:4639 / 4677
页数:39
相关论文
共 50 条
  • [1] Variance reduction on general adaptive stochastic mirror descent
    Wenjie Li
    Zhanyu Wang
    Yichen Zhang
    Guang Cheng
    [J]. Machine Learning, 2022, 111 : 4639 - 4677
  • [2] On stochastic mirror descent with interacting particles: Convergence properties and variance reduction
    Borovykh, A.
    Kantas, N.
    Parpas, P.
    Pavliotis, G. A.
    [J]. PHYSICA D-NONLINEAR PHENOMENA, 2021, 418
  • [3] Continuous-Time Stochastic Mirror Descent on a Network: Variance Reduction, Consensus, Convergence
    Raginsky, Maxim
    Bouvrie, Jake
    [J]. 2012 IEEE 51ST ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2012, : 6793 - 6800
  • [4] Stochastic gradient descent with variance reduction technique
    Zhang, Jinjing
    Hu, Fei
    Xu, Xiaofei
    Li, Li
    [J]. WEB INTELLIGENCE, 2018, 16 (03) : 187 - 194
  • [5] Adaptive Stochastic Mirror Descent for Constrained Optimization
    Bayandina, Anastasia
    [J]. 2017 CONSTRUCTIVE NONSMOOTH ANALYSIS AND RELATED TOPICS (DEDICATED TO THE MEMORY OF V.F. DEMYANOV) (CNSA), 2017, : 40 - 43
  • [6] Distributed and asynchronous Stochastic Gradient Descent with variance reduction
    Ming, Yuewei
    Zhao, Yawei
    Wu, Chengkun
    Li, Kuan
    Yin, Jianping
    [J]. NEUROCOMPUTING, 2018, 281 : 27 - 36
  • [7] On Variance Reduction in Stochastic Gradient Descent and its Asynchronous Variants
    Reddi, Sashank J.
    Hefny, Ahmed
    Sra, Suvrit
    Poczos, Barnabas
    Smola, Alex
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 28 (NIPS 2015), 2015, 28
  • [8] Adaptive Importance Sampling meets Mirror Descent: a Bias-Variance Tradeoff
    Korba, Anna
    Portier, Francois
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [9] Adaptive optics system with micromachined mirror array and stochastic gradient descent controller
    Weyrauch, T
    Vorontsov, MA
    Bifano, TG
    Giles, MK
    [J]. HIGH-RESOLUTION WAVEFRONT CONTROL: METHODS, DEVICES, AND APPLICATIONS II, 2000, 4124 : 178 - 188
  • [10] An Accelerated Stochastic Mirror Descent Method
    Jiang, Bo-Ou
    Yuan, Ya-Xiang
    [J]. JOURNAL OF THE OPERATIONS RESEARCH SOCIETY OF CHINA, 2024, 12 (03) : 549 - 571