Variance reduction on general adaptive stochastic mirror descent

被引:1
|
作者
Li, Wenjie [1 ]
Wang, Zhanyu [1 ]
Zhang, Yichen [2 ]
Cheng, Guang [3 ]
机构
[1] Purdue Univ, Dept Stat, W Lafayette, IN 47907 USA
[2] Purdue Univ, Krannert Sch Management, W Lafayette, IN 47907 USA
[3] Univ Calif Los Angeles, Dept Stat, Los Angeles, CA USA
关键词
Variance reduction; Adaptive mirror descent; Nonconvex nonsmooth optimization; General framework; Convergence analysis;
D O I
10.1007/s10994-022-06227-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this work, we propose a simple and generalized algorithmic framework for applying variance reduction to adaptive mirror descent algorithms for faster convergence. We introduce the SVRAMD algorithm, and provide its general convergence analysis in both the nonsmooth nonconvex optimization problem and the generalized P-L conditioned nonconvex optimization problem. We prove that variance reduction can reduce the gradient complexity of all adaptive mirror descent algorithms that satisfy a mild assumption and thus accelerate their convergence. In particular, our general theory implies that variance reduction can be applied to different algorithms with their distinct choices of the proximal function, such as gradient descent with time-varying step sizes, mirror descent with L-1 mirror maps, and self-adaptive algorithms such as AdaGrad and RMSProp. Moreover, the proved convergence rates of SVRAMD recover the existing rates without complicated algorithmic components, which indicates their optimality. Extensive experiments validate our theoretical findings.
引用
下载
收藏
页码:4639 / 4677
页数:39
相关论文
共 50 条
  • [31] Stochastic incremental mirror descent algorithms with Nesterov smoothing
    Sandy Bitterlich
    Sorin-Mihai Grad
    Numerical Algorithms, 2024, 95 : 351 - 382
  • [32] Validation analysis of mirror descent stochastic approximation method
    Lan, Guanghui
    Nemirovski, Arkadi
    Shapiro, Alexander
    MATHEMATICAL PROGRAMMING, 2012, 134 (02) : 425 - 458
  • [33] General inertial proximal stochastic variance reduction gradient for nonconvex nonsmooth optimization
    Shuya Sun
    Lulu He
    Journal of Inequalities and Applications, 2023
  • [34] General inertial proximal stochastic variance reduction gradient for nonconvex nonsmooth optimization
    Sun, Shuya
    He, Lulu
    JOURNAL OF INEQUALITIES AND APPLICATIONS, 2023, 2023 (01)
  • [35] Large-scale machine learning with synchronous parallel adaptive stochastic variance reduction gradient descent for high-dimensional blindness detection on spark
    Chuandong Qin
    Yiqing Zhang
    Yu Cao
    The Journal of Supercomputing, 81 (4)
  • [36] Inexact stochastic mirror descent for two-stage nonlinear stochastic programs
    Guigues, Vincent
    MATHEMATICAL PROGRAMMING, 2021, 187 (1-2) : 533 - 577
  • [37] Inexact stochastic mirror descent for two-stage nonlinear stochastic programs
    Vincent Guigues
    Mathematical Programming, 2021, 187 : 533 - 577
  • [38] Batching Adaptive Variance Reduction
    Song, Chenxiao
    Kawai, Reiichiro
    ACM TRANSACTIONS ON MODELING AND COMPUTER SIMULATION, 2023, 33 (1-2):
  • [39] Adaptive Mirror Descent for the Network Utility Maximization Problem
    Ivanova, Anastasiya
    Stonyakin, Fedor
    Pasechnyuk, Dmitry
    Vorontsova, Evgeniya
    Gasnikov, Alexander
    IFAC PAPERSONLINE, 2020, 53 (02): : 7851 - 7856
  • [40] Variance-Reduced Stochastic Gradient Descent on Streaming Data
    Jothimurugesan, Ellango
    Tahmasbi, Ashraf
    Gibbons, Phillip B.
    Tirthapura, Srikanta
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31