Improved Zeroth-Order Variance Reduced Algorithms and Analysis for Nonconvex Optimization

被引:0
|
作者
Ji, Kaiyi [1 ]
Wang, Zhe [1 ]
Zhou, Yi [2 ]
Liang, Yingbin [1 ]
机构
[1] Ohio State Univ, Dept Elect & Comp Engn, Columbus, OH 43210 USA
[2] Duke Univ, Dept Elect & Comp Engn, Durham, NC 27706 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Two types of zeroth-order stochastic algorithms have recently been designed for nonconvex optimization respectively based on the first-order techniques SVRG and SARAH/SPIDER. This paper addresses several important issues that are still open in these methods. First, all existing SVRG-type zeroth-order algorithms suffer from worse function query complexities than either zerothorder gradient descent (ZO-GD) or stochastic gradient descent (ZO-SGD). In this paper, we propose a new algorithm ZO-SVRG-Coord-Rand and develop a new analysis for an existing ZO-SVRG-Coord algorithm proposed in Liu et al. 2018b, and show that both ZO-SVRG-Coord-Rand and ZO-SVRG-Coord (under our new analysis) outperform other exiting SVRG-type zeroth-order methods as well as ZO-GD and ZO-SGD. Second, the existing SPIDER-type algorithm SPIDER-SZO (Fang et al., 2018) has superior theoretical performance, but suffers from the generation of a large number of Gaussian random variables as well as root epsilon-level stepsize in practice. In this paper, we develop a new algorithm ZO-SPIDER-Coord, which is free from Gaussian variable generation and allows a large constant stepsize while maintaining the same convergence rate and query complexity, and we further show that ZO-SPIDER-Coord automatically achieves a linear convergence rate as the iterate enters into a local PL region without restart and algorithmic modification.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] Certified Multifidelity Zeroth-Order Optimization
    de Montbrun, Etienne
    Gerchinovitz, Sebastien
    [J]. SIAM-ASA Journal on Uncertainty Quantification, 2024, 12 (04): : 1135 - 1164
  • [22] Zeroth-Order Nonconvex Stochastic Optimization: Handling Constraints, High Dimensionality, and Saddle Points
    Krishnakumar Balasubramanian
    Saeed Ghadimi
    [J]. Foundations of Computational Mathematics, 2022, 22 : 35 - 76
  • [23] Zeroth-Order Alternating Gradient Descent Ascent Algorithms for A Class of Nonconvex-Nonconcave Minimax Problems
    Xu, Zi
    Wang, Zi-Qi
    Wang, Jun-Lin
    Dai, Yu-Hong
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [24] UNBIASED GRADIENT SIMULATION FOR ZEROTH-ORDER OPTIMIZATION
    Chen, Guanting
    [J]. 2020 WINTER SIMULATION CONFERENCE (WSC), 2020, : 2947 - 2959
  • [25] Stochastic Zeroth-order Optimization in High Dimensions
    Wang, Yining
    Du, Simon S.
    Balakrishnan, Sivaraman
    Singh, Aarti
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 84, 2018, 84
  • [26] STOCHASTIC FIRST- AND ZEROTH-ORDER METHODS FOR NONCONVEX STOCHASTIC PROGRAMMING
    Ghadimi, Saeed
    Lan, Guanghui
    [J]. SIAM JOURNAL ON OPTIMIZATION, 2013, 23 (04) : 2341 - 2368
  • [27] Zeroth-order optimization with orthogonal random directions
    David Kozak
    Cesare Molinari
    Lorenzo Rosasco
    Luis Tenorio
    Silvia Villa
    [J]. Mathematical Programming, 2023, 199 : 1179 - 1219
  • [28] Zeroth-order Optimization with Weak Dimension Dependency
    Yue, Pengyun
    Yang, Long
    Fang, Cong
    Lin, Zhouchen
    [J]. THIRTY SIXTH ANNUAL CONFERENCE ON LEARNING THEORY, VOL 195, 2023, 195
  • [29] Zeroth-order optimization with orthogonal random directions
    Kozak, David
    Molinari, Cesare
    Rosasco, Lorenzo
    Tenorio, Luis
    Villa, Silvia
    [J]. MATHEMATICAL PROGRAMMING, 2023, 199 (1-2) : 1179 - 1219
  • [30] ZEROTH-ORDER RIEMANNIAN AVERAGING STOCHASTIC APPROXIMATION ALGORITHMS
    Li, Jiaxiang
    Balasubramanian, Krishnakumar
    Ma, Shiqian
    [J]. SIAM Journal on Optimization, 2024, 34 (04) : 3314 - 3341