ZEROTH-ORDER REGULARIZED OPTIMIZATION (ZORO): APPROXIMATELY SPARSE GRADIENTS AND ADAPTIVE SAMPLING

被引:7
|
作者
Cai, HanQin [1 ]
McKenzie, Daniel [1 ]
Yin, Wotao [2 ]
Zhang, Zhenliang [3 ]
机构
[1] Univ Calif Los Angeles, Dept Math, Los Angeles, CA 90095 USA
[2] Alibaba US, Damo Acad, Bellevue, WA 98004 USA
[3] Xmotors AI, Mountain View, CA 94043 USA
关键词
zeroth-order optimization; black-box optimization; derivative-free optimization; regularized optimization; sparse gradients; sparse adversarial attack; CONVERGENCE; PARAMETER; SEARCH;
D O I
10.1137/21M1392966
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
We consider the problem of minimizing a high-dimensional objective function, which may include a regularization term, using only (possibly noisy) evaluations of the function. Such optimization is also called derivative-free, zeroth-order, or black-box optimization. We propose a new zeroth-order regularized optimization method, dubbed ZORO. When the underlying gradient is approximately sparse at an iterate, ZORO needs very few objective function evaluations to obtain a new iterate that decreases the objective function. We achieve this with an adaptive, randomized gradient estimator, followed by an inexact proximal-gradient scheme. Under a novel approximately sparse gradient assumption and various different convex settings, we show that the (theoretical and empirical) convergence rate of ZORO is only logarithmically dependent on the problem dimension. Numerical experiments show that ZORO outperforms existing methods with similar assumptions, on both synthetic and real datasets.
引用
收藏
页码:687 / 714
页数:28
相关论文
共 50 条
  • [21] Zeroth-Order Method for Distributed Optimization With Approximate Projections
    Yuan, Deming
    Ho, Daniel W. C.
    Xu, Shengyuan
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2016, 27 (02) : 284 - 294
  • [22] Zeroth-Order Stochastic Variance Reduction for Nonconvex Optimization
    Liu, Sijia
    Kailkhura, Bhavya
    Chen, Pin-Yu
    Ting, Paishun
    Chang, Shiyu
    Amini, Lisa
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [23] Stochastic Zeroth-Order Optimization under Nonstationarity and Nonconvexity
    Roy, Abhishek
    Balasubramanian, Krishnakumar
    Ghadimi, Saeed
    Mohapatra, Prasant
    JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [24] Distributed zeroth-order online optimization with communication delays
    Hayashi, Naoki (n.hayashi@sys.es.osaka-u.ac.jp), 2025, 19 (01):
  • [25] Zeroth-Order Optimization for Composite Problems with Functional Constraints
    Li, Zichong
    Chen, Pin-Yu
    Liu, Sijia
    Lu, Songtao
    Xu, Yangyang
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 7453 - 7461
  • [26] Safe Zeroth-Order Optimization Using Linear Programs
    Guo, Baiwei
    Wang, Yang
    Jiang, Yuning
    Kamgarpour, Maryam
    Ferrari-Trecate, Giancarlo
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 556 - 561
  • [27] Zeroth-order Gradient Tracking for Distributed Constrained Optimization
    Cheng, Songsong
    Yu, Xin
    Fan, Yuan
    Xiao, Gaoxi
    IFAC PAPERSONLINE, 2023, 56 (02): : 5197 - 5202
  • [28] ZO-AdaMM: Zeroth-Order Adaptive Momentum Method for Black-Box Optimization
    Chen, Xiangyi
    Liu, Sijia
    Xu, Kaidi
    Li, Xingguo
    Lin, Xue
    Hong, Mingyi
    Cox, David
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [29] Safe zeroth-order optimization using quadratic local approximations
    Guo, Baiwei
    Jiang, Yuning
    Ferrari-Trecate, Giancarlo
    Kamgarpour, Maryam
    Automatica, 2025, 174
  • [30] Sequential stochastic blackbox optimization with zeroth-order gradient estimators
    Audet, Charles
    Bigeon, Jean
    Couderc, Romain
    Kokkolaras, Michael
    AIMS MATHEMATICS, 2023, 8 (11): : 25922 - 25956