ZEROTH-ORDER REGULARIZED OPTIMIZATION (ZORO): APPROXIMATELY SPARSE GRADIENTS AND ADAPTIVE SAMPLING

被引:7
|
作者
Cai, HanQin [1 ]
McKenzie, Daniel [1 ]
Yin, Wotao [2 ]
Zhang, Zhenliang [3 ]
机构
[1] Univ Calif Los Angeles, Dept Math, Los Angeles, CA 90095 USA
[2] Alibaba US, Damo Acad, Bellevue, WA 98004 USA
[3] Xmotors AI, Mountain View, CA 94043 USA
关键词
zeroth-order optimization; black-box optimization; derivative-free optimization; regularized optimization; sparse gradients; sparse adversarial attack; CONVERGENCE; PARAMETER; SEARCH;
D O I
10.1137/21M1392966
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
We consider the problem of minimizing a high-dimensional objective function, which may include a regularization term, using only (possibly noisy) evaluations of the function. Such optimization is also called derivative-free, zeroth-order, or black-box optimization. We propose a new zeroth-order regularized optimization method, dubbed ZORO. When the underlying gradient is approximately sparse at an iterate, ZORO needs very few objective function evaluations to obtain a new iterate that decreases the objective function. We achieve this with an adaptive, randomized gradient estimator, followed by an inexact proximal-gradient scheme. Under a novel approximately sparse gradient assumption and various different convex settings, we show that the (theoretical and empirical) convergence rate of ZORO is only logarithmically dependent on the problem dimension. Numerical experiments show that ZORO outperforms existing methods with similar assumptions, on both synthetic and real datasets.
引用
收藏
页码:687 / 714
页数:28
相关论文
共 50 条
  • [11] Zeroth-order optimization with orthogonal random directions
    Kozak, David
    Molinari, Cesare
    Rosasco, Lorenzo
    Tenorio, Luis
    Villa, Silvia
    MATHEMATICAL PROGRAMMING, 2023, 199 (1-2) : 1179 - 1219
  • [12] Zeroth-order Optimization with Weak Dimension Dependency
    Yue, Pengyun
    Yang, Long
    Fang, Cong
    Lin, Zhouchen
    THIRTY SIXTH ANNUAL CONFERENCE ON LEARNING THEORY, VOL 195, 2023, 195
  • [13] Zeroth-order optimization with orthogonal random directions
    David Kozak
    Cesare Molinari
    Lorenzo Rosasco
    Luis Tenorio
    Silvia Villa
    Mathematical Programming, 2023, 199 : 1179 - 1219
  • [14] Zeroth-Order Diffusion Adaptive Filter Over Networks
    Zhang, Mengfei
    Jin, Danqi
    Chen, Jie
    Ni, Jingen
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2021, 69 : 589 - 602
  • [15] Adaptive Zeroth-Order Optimisation of Nonconvex Composite Objectives
    Shao, Weijia
    Albayrak, Sahin
    MACHINE LEARNING, OPTIMIZATION, AND DATA SCIENCE, LOD 2022, PT I, 2023, 13810 : 573 - 595
  • [16] Asynchronous Zeroth-Order Distributed Optimization with Residual Feedback
    Shen, Yi
    Zhang, Yan
    Nivison, Scott
    Bell, Zachary, I
    Zavlanos, Michael M.
    2021 60TH IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2021, : 3349 - 3354
  • [17] SMALL ERRORS IN RANDOM ZEROTH-ORDER OPTIMIZATION ARE IMAGINARY
    Jongeneel, Wouter
    Yue, Man-Chung
    Kuhn, Daniel
    SIAM Journal on Optimization, 2024, 34 (03) : 2638 - 2670
  • [18] Stochastic Zeroth-Order Riemannian Derivative Estimation and Optimization
    Li, Jiaxiang
    Balasubramanian, Krishnakumar
    Ma, Shiqian
    MATHEMATICS OF OPERATIONS RESEARCH, 2023, 48 (02) : 1183 - 1211
  • [19] Zeroth-order algorithms for stochastic distributed nonconvex optimization
    Yi, Xinlei
    Zhang, Shengjun
    Yang, Tao
    Johansson, Karl H.
    AUTOMATICA, 2022, 142
  • [20] Stochastic Zeroth-Order Optimization under Nonstationarity and Nonconvexity
    Roy, Abhishek
    Balasubramanian, Krishnakumar
    Ghadimi, Saeed
    Mohapatra, Prasant
    Journal of Machine Learning Research, 2022, 23