Gradient-free methods for non-smooth convex stochastic optimization with heavy-tailed noise on convex compact

被引:0
|
作者
Nikita Kornilov
Alexander Gasnikov
Pavel Dvurechensky
Darina Dvinskikh
机构
[1] Moscow Institute of Physics and Technology,
[2] Weierstrass Institute for Applied Analysis and Stochastics,undefined
[3] HSE University,undefined
[4] Skoltech,undefined
[5] ISP RAS Research Center for Trusted Artificial Intelligence,undefined
来源
关键词
Zeroth-order optimization; Derivative-free optimization; Stochastic optimization; Non-smooth problems; Heavy tails; Gradient clipping; Stochastic mirror descent;
D O I
暂无
中图分类号
学科分类号
摘要
We present two easy-to-implement gradient-free/zeroth-order methods to optimize a stochastic non-smooth function accessible only via a black-box. The methods are built upon efficient first-order methods in the heavy-tailed case, i.e., when the gradient noise has infinite variance but bounded (1+κ)\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$(1+\kappa)$$\end{document}-th moment for some κ∈(0,1]\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\kappa \in(0,1]$$\end{document}. The first algorithm is based on the stochastic mirror descent with a particular class of uniformly convex mirror maps which is robust to heavy-tailed noise. The second algorithm is based on the stochastic mirror descent and gradient clipping technique. Additionally, for the objective functions satisfying the r-growth condition, faster algorithms are proposed based on these methods and the restart technique.
引用
收藏
相关论文
共 50 条
  • [1] Gradient-free methods for non-smooth convex stochastic optimization with heavy-tailed noise on convex compact
    Kornilov, Nikita
    Gasnikov, Alexander
    Dvurechensky, Pavel
    Dvinskikh, Darina
    COMPUTATIONAL MANAGEMENT SCIENCE, 2023, 20 (01)
  • [2] Decentralized Gradient-Free Methods for Stochastic Non-smooth Non-convex Optimization
    Lin, Zhenwei
    Xia, Jingfan
    Deng, Qi
    Luo, Luo
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 16, 2024, : 17477 - 17486
  • [3] High-Probability Complexity Bounds for Non-smooth Stochastic Convex Optimization with Heavy-Tailed Noise
    Gorbunov, Eduard
    Danilova, Marina
    Shibaev, Innokentiy
    Dvurechensky, Pavel
    Gasnikov, Alexander
    JOURNAL OF OPTIMIZATION THEORY AND APPLICATIONS, 2024, 203 (03) : 2679 - 2738
  • [4] Simple Stochastic Gradient Methods for Non-Smooth Non-Convex Regularized Optimization
    Metel, Michael R.
    Takeda, Akiko
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [5] Gradient-free Federated Learning Methods with l1 and l2-randomization for Non-smooth Convex Stochastic Optimization Problems
    B. A. Alashqar
    A. V. Gasnikov
    D. M. Dvinskikh
    A. V. Lobanov
    Computational Mathematics and Mathematical Physics, 2023, 63 : 1600 - 1653
  • [6] Gradient-free Federated Learning Methods with l1 and l2-randomization for Non-smooth Convex Stochastic Optimization Problems
    Alashqar, B. A.
    Gasnikov, A. V.
    Dvinskikh, D. M.
    Lobanov, A. V.
    COMPUTATIONAL MATHEMATICS AND MATHEMATICAL PHYSICS, 2023, 63 (09) : 1600 - 1653
  • [7] Inexact Proximal Gradient Methods for Non-Convex and Non-Smooth Optimization
    Gu, Bin
    Wang, De
    Huo, Zhouyuan
    Huang, Heng
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 3093 - 3100
  • [8] Breaking the Lower Bound with (Little) Structure: Acceleration in Non-Convex Stochastic Optimization with Heavy-Tailed Noise
    Liu, Zijian
    Zhang, Jiawei
    Zhou, Zhengyuan
    THIRTY SIXTH ANNUAL CONFERENCE ON LEARNING THEORY, VOL 195, 2023, 195
  • [9] A memory gradient method for non-smooth convex optimization
    Ou, Yigui
    Liu, Yuanwen
    INTERNATIONAL JOURNAL OF COMPUTER MATHEMATICS, 2015, 92 (08) : 1625 - 1642
  • [10] Distributed Stochastic Strongly Convex Optimization under Heavy-Tailed Noises
    Sun, Chao
    Chen, Bo
    2024 IEEE INTERNATIONAL CONFERENCE ON CYBERNETICS AND INTELLIGENT SYSTEMS, CIS AND IEEE INTERNATIONAL CONFERENCE ON ROBOTICS, AUTOMATION AND MECHATRONICS, RAM, CIS-RAM 2024, 2024, : 150 - 155