Accelerated Zeroth-order Method for Non-Smooth Stochastic Convex Optimization Problem with Infinite Variance

被引:0
|
作者
Kornilov, Nikita [1 ,2 ]
Shamir, Ohad [3 ]
Lobanov, Aleksandr [1 ,4 ]
Dvinskikh, Darina [4 ,5 ]
Gasnikov, Alexander [1 ,2 ,4 ]
Shibaev, Innokentiy [1 ,6 ]
Gorbunov, Eduard [7 ]
Horvath, Samuel [7 ]
机构
[1] MIPT, Dolgoprudnyi, Russia
[2] SkolTech, Moscow, Russia
[3] Weizmann Inst Sci, Rehovot, Israel
[4] RAS, ISP, Moscow, Russia
[5] HSE Univ, Moscow, Russia
[6] RAS, IITP, Moscow, Russia
[7] MBZUAI, Abu Dhabi, U Arab Emirates
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we consider non-smooth stochastic convex optimization with two function evaluations per round under infinite noise variance. In the classical setting when noise has finite variance, an optimal algorithm, built upon the batched accelerated gradient method, was proposed in [17]. This optimality is defined in terms of iteration and oracle complexity, as well as the maximal admissible level of adversarial noise. However, the assumption of finite variance is burdensome and it might not hold in many practical scenarios. To address this, we demonstrate how to adapt a refined clipped version of the accelerated gradient (Stochastic Similar Triangles) method from [35] for a two-point zero-order oracle. This adaptation entails extending the batching technique to accommodate infinite variance - a non-trivial task that stands as a distinct contribution of this paper.
引用
收藏
页数:20
相关论文
共 50 条
  • [31] Non-smooth setting of stochastic decentralized convex optimization problem over time-varying Graphs
    Lobanov, Aleksandr
    Veprikov, Andrew
    Konin, Georgiy
    Beznosikov, Aleksandr
    Gasnikov, Alexander
    Kovalev, Dmitry
    COMPUTATIONAL MANAGEMENT SCIENCE, 2023, 20 (01)
  • [32] Non-smooth setting of stochastic decentralized convex optimization problem over time-varying Graphs
    Aleksandr Lobanov
    Andrew Veprikov
    Georgiy Konin
    Aleksandr Beznosikov
    Alexander Gasnikov
    Dmitry Kovalev
    Computational Management Science, 2023, 20
  • [33] Communication-Efficient Stochastic Zeroth-Order Optimization for Federated Learning
    Fang, Wenzhi
    Yu, Ziyi
    Jiang, Yuning
    Shi, Yuanming
    Jones, Colin N.
    Zhou, Yong
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2022, 70 : 5058 - 5073
  • [34] Simple Stochastic Gradient Methods for Non-Smooth Non-Convex Regularized Optimization
    Metel, Michael R.
    Takeda, Akiko
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [35] Stochastic Proximal Methods for Non-Smooth Non-Convex Constrained Sparse Optimization
    Metel, Michael R.
    Takeda, Akiko
    JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22
  • [36] Optimal, Stochastic, Non-smooth, Non-convex Optimization through Online-to-Non-convex Conversion
    Cutkosky, Ashok
    Mehta, Harsh
    Orabona, Francesco
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202
  • [37] Stochastic proximal methods for non-smooth non-convex constrained sparse optimization
    Metel, Michael R.
    Takeda, Akiko
    Journal of Machine Learning Research, 2021, 22
  • [38] Escaping Saddle Points for Zeroth-order Non-convex Optimization using Estimated Gradient Descent
    Bai, Qinbo
    Agarwal, Mridul
    Aggarwal, Vaneet
    2020 54TH ANNUAL CONFERENCE ON INFORMATION SCIENCES AND SYSTEMS (CISS), 2020, : 132 - 137
  • [39] Stochastic zeroth-order gradient and Hessian estimators: variance reduction and refined bias bounds
    Feng, Yasong
    Wang, Tianyu
    INFORMATION AND INFERENCE-A JOURNAL OF THE IMA, 2023, 12 (03)
  • [40] AN ACCELERATED METHOD FOR DERIVATIVE-FREE SMOOTH STOCHASTIC CONVEX OPTIMIZATION
    Gorbunov, Eduard
    Dvurechensky, Pavel
    Gasnikov, Alexander
    SIAM JOURNAL ON OPTIMIZATION, 2022, 32 (02) : 1210 - 1238