ZEROTH-ORDER REGULARIZED OPTIMIZATION (ZORO): APPROXIMATELY SPARSE GRADIENTS AND ADAPTIVE SAMPLING

被引:7
|
作者
Cai, HanQin [1 ]
McKenzie, Daniel [1 ]
Yin, Wotao [2 ]
Zhang, Zhenliang [3 ]
机构
[1] Univ Calif Los Angeles, Dept Math, Los Angeles, CA 90095 USA
[2] Alibaba US, Damo Acad, Bellevue, WA 98004 USA
[3] Xmotors AI, Mountain View, CA 94043 USA
关键词
zeroth-order optimization; black-box optimization; derivative-free optimization; regularized optimization; sparse gradients; sparse adversarial attack; CONVERGENCE; PARAMETER; SEARCH;
D O I
10.1137/21M1392966
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
We consider the problem of minimizing a high-dimensional objective function, which may include a regularization term, using only (possibly noisy) evaluations of the function. Such optimization is also called derivative-free, zeroth-order, or black-box optimization. We propose a new zeroth-order regularized optimization method, dubbed ZORO. When the underlying gradient is approximately sparse at an iterate, ZORO needs very few objective function evaluations to obtain a new iterate that decreases the objective function. We achieve this with an adaptive, randomized gradient estimator, followed by an inexact proximal-gradient scheme. Under a novel approximately sparse gradient assumption and various different convex settings, we show that the (theoretical and empirical) convergence rate of ZORO is only logarithmically dependent on the problem dimension. Numerical experiments show that ZORO outperforms existing methods with similar assumptions, on both synthetic and real datasets.
引用
收藏
页码:687 / 714
页数:28
相关论文
共 50 条
  • [1] Adaptive sampling quasi-Newton methods for zeroth-order stochastic optimization
    Bollapragada, Raghu
    Wild, Stefan M. M.
    MATHEMATICAL PROGRAMMING COMPUTATION, 2023, 15 (02) : 327 - 364
  • [2] Adaptive sampling quasi-Newton methods for zeroth-order stochastic optimization
    Raghu Bollapragada
    Stefan M. Wild
    Mathematical Programming Computation, 2023, 15 : 327 - 364
  • [3] Adaptive Evolution Strategies for Stochastic Zeroth-Order Optimization
    He, Xiaoyu
    Zheng, Zibin
    Chen, Zefeng
    Zhou, Yuren
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2022, 6 (05): : 1271 - 1285
  • [4] A ZEROTH-ORDER LEARNING ALGORITHM FOR ERGODIC OPTIMIZATION OF WIRELESS SYSTEMS WITH NO MODELS AND NO GRADIENTS
    Kalogerias, Dionysios S.
    Eisen, Mark
    Pappas, George J.
    Ribeiro, Alejandro
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 5045 - 5049
  • [5] Learning Sampling Policy to Achieve Fewer Queries for Zeroth-Order Optimization
    Zhai, Zhou
    Shi, Wanli
    Huang, Heng
    Chang, Yi
    Gu, Bin
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [6] Certified Multifidelity Zeroth-Order Optimization
    de Montbrun, Etienne
    Gerchinovitz, Sebastien
    SIAM-ASA Journal on Uncertainty Quantification, 2024, 12 (04): : 1135 - 1164
  • [7] Communication-Efficient Zeroth-Order Adaptive Optimization for Federated Learning
    Xie, Ping
    Gao, Xiangrui
    Li, Fan
    Xing, Ling
    Zhang, Yu
    Sun, Hanxiao
    MATHEMATICS, 2024, 12 (08)
  • [8] Stochastic Zeroth-order Optimization in High Dimensions
    Wang, Yining
    Du, Simon S.
    Balakrishnan, Sivaraman
    Singh, Aarti
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 84, 2018, 84
  • [9] UNBIASED GRADIENT SIMULATION FOR ZEROTH-ORDER OPTIMIZATION
    Chen, Guanting
    2020 WINTER SIMULATION CONFERENCE (WSC), 2020, : 2947 - 2959
  • [10] Adaptive and Communication-Efficient Zeroth-Order Optimization for Distributed Internet of Things
    Dang, Qianlong
    Yang, Shuai
    Liu, Qiqi
    Ruan, Junhu
    IEEE Internet of Things Journal, 2024, 11 (22) : 37200 - 37213