Stochastic Recursive Variance-Reduced Cubic Regularization Methods

被引:0
|
作者
Zhou, Dongruo [1 ]
Gu, Quanquan [1 ]
机构
[1] Univ Calif Los Angeles, Dept Comp Sci, Los Angeles, CA 90024 USA
基金
美国国家科学基金会;
关键词
COMPLEXITY;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Stochastic Variance-Reduced Cubic regularization (SVRC) algorithms have received increasing attention due to its improved gradient/Hessian complexities (i.e., number of queries to stochastic gradient/Hessian oracles) to find local minima for nonconvex finite-sum optimization. However, it is unclear whether existing SVRC algorithms can be further improved. Moreover, the semi-stochastic Hessian estimator adopted in existing SVRC algorithms prevents the use of Hessian-vector product-based fast cubic subproblem solvers, which makes SVRC algorithms computationally intractable for high-dimensional problems. In this paper, we first present a Stochastic Recursive Variance-Reduced Cubic regularization method (SRVRC) using a recursively updated semi-stochastic gradient and Hessian estimators. It enjoys improved gradient and Hessian complexities to find an (epsilon, root epsilon)-approximate local minimum, and outperforms the state-of-the-art SVRC algorithms. Built upon SRVRC, we further propose a Hessian-free SRVRC algorithm, namely SRVRC free, which only needs (O) over tilde (n epsilon(-2) boolean AND epsilon(-3)) stochastic gradient and Hessian-vector product computations, where n is the number of component functions in the finite-sum objective and epsilon is the optimization precision. This outperforms the best-known result (O) over tilde(epsilon(-3.5)) achieved by stochastic cubic regularization algorithm proposed in Tripuraileili et al. (2018).
引用
收藏
页码:3980 / 3989
页数:10
相关论文
共 50 条
  • [1] Stochastic Variance-Reduced Cubic Regularization Methods
    Zhou, Dongruo
    Xu, Pan
    Gu, Quanquan
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2019, 20
  • [2] Stochastic variance-reduced cubic regularization methods
    Zhou, Dongruo
    Xu, Pan
    Gu, Quanquan
    [J]. Journal of Machine Learning Research, 2019, 20
  • [3] Stochastic Variance-Reduced Cubic Regularization for Nonconvex Optimization
    Wang, Zhe
    Zhou, Yi
    Liang, Yingbin
    Lan, Guanghui
    [J]. 22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [4] Stochastic Variance-Reduced Cubic Regularized Newton Methods
    Zhou, Dongruo
    Xu, Pan
    Gu, Quanquan
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [5] Accelerating variance-reduced stochastic gradient methods
    Derek Driggs
    Matthias J. Ehrhardt
    Carola-Bibiane Schönlieb
    [J]. Mathematical Programming, 2022, 191 : 671 - 715
  • [6] Accelerating variance-reduced stochastic gradient methods
    Driggs, Derek
    Ehrhardt, Matthias J.
    Schonlieb, Carola-Bibiane
    [J]. MATHEMATICAL PROGRAMMING, 2022, 191 (02) : 671 - 715
  • [7] Stochastic Variance-Reduced Hamilton Monte Carlo Methods
    Zou, Difan
    Xu, Pan
    Gu, Quanquan
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [8] Cocoercivity, smoothness and bias in variance-reduced stochastic gradient methods
    Martin Morin
    Pontus Giselsson
    [J]. Numerical Algorithms, 2022, 91 : 749 - 772
  • [9] Cocoercivity, smoothness and bias in variance-reduced stochastic gradient methods
    Morin, Martin
    Giselsson, Pontus
    [J]. NUMERICAL ALGORITHMS, 2022, 91 (02) : 749 - 772
  • [10] Stochastic Variance-Reduced Policy Gradient
    Papini, Matteo
    Binaghi, Damiano
    Canonaco, Giuseppe
    Pirotta, Matteo
    Restelli, Marcello
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80