ZEROTH-ORDER REGULARIZED OPTIMIZATION (ZORO): APPROXIMATELY SPARSE GRADIENTS AND ADAPTIVE SAMPLING

被引:7
|
作者
Cai, HanQin [1 ]
McKenzie, Daniel [1 ]
Yin, Wotao [2 ]
Zhang, Zhenliang [3 ]
机构
[1] Univ Calif Los Angeles, Dept Math, Los Angeles, CA 90095 USA
[2] Alibaba US, Damo Acad, Bellevue, WA 98004 USA
[3] Xmotors AI, Mountain View, CA 94043 USA
关键词
zeroth-order optimization; black-box optimization; derivative-free optimization; regularized optimization; sparse gradients; sparse adversarial attack; CONVERGENCE; PARAMETER; SEARCH;
D O I
10.1137/21M1392966
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
We consider the problem of minimizing a high-dimensional objective function, which may include a regularization term, using only (possibly noisy) evaluations of the function. Such optimization is also called derivative-free, zeroth-order, or black-box optimization. We propose a new zeroth-order regularized optimization method, dubbed ZORO. When the underlying gradient is approximately sparse at an iterate, ZORO needs very few objective function evaluations to obtain a new iterate that decreases the objective function. We achieve this with an adaptive, randomized gradient estimator, followed by an inexact proximal-gradient scheme. Under a novel approximately sparse gradient assumption and various different convex settings, we show that the (theoretical and empirical) convergence rate of ZORO is only logarithmically dependent on the problem dimension. Numerical experiments show that ZORO outperforms existing methods with similar assumptions, on both synthetic and real datasets.
引用
收藏
页码:687 / 714
页数:28
相关论文
共 50 条
  • [41] ZONE: Zeroth-Order Nonconvex Multiagent Optimization Over Networks
    Hajinezhad, Davood
    Hong, Mingyi
    Garcia, Alfredo
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2019, 64 (10) : 3995 - 4010
  • [42] Gradients in the ab initio scalar zeroth-order regular approximation (ZORA) approach
    van Lenthe, JH
    Faas, S
    Snijders, JG
    CHEMICAL PHYSICS LETTERS, 2000, 328 (1-2) : 107 - 112
  • [43] Lazy Queries Can Reduce Variance in Zeroth-Order Optimization
    Xiao, Quan
    Ling, Qing
    Chen, Tianyi
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2023, 71 : 3695 - 3709
  • [44] Privacy-Preserved Distributed Learning With Zeroth-Order Optimization
    Gratton, Cristiano
    Venkategowda, Naveen K. D.
    Arablouei, Reza
    Werner, Stefan
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2022, 17 : 265 - 279
  • [45] A Parallel Zeroth-Order Framework for Efficient Cellular Network Optimization
    He, Pengcheng
    Lu, Siyuan
    Xu, Fan
    Kang, Yibin
    Yan, Qi
    Shi, Qingjiang
    IEEE Transactions on Wireless Communications, 2024, 23 (11) : 17522 - 17538
  • [46] Improved Zeroth-Order Variance Reduced Algorithms and Analysis for Nonconvex Optimization
    Ji, Kaiyi
    Wang, Zhe
    Zhou, Yi
    Liang, Yingbin
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97
  • [47] Contrasting Exploration in Parameter and Action Space: A Zeroth-Order Optimization Perspective
    Vemula, Anirudh
    Sun, Wen
    Bagnell, J. Andrew
    22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [48] Safe Zeroth-Order Convex Optimization Using Quadratic Local Approximations
    Guo, Baiwei
    Jiang, Yuning
    Kamgarpour, Maryam
    Ferrari-Trecate, Giancarlo
    2023 EUROPEAN CONTROL CONFERENCE, ECC, 2023,
  • [49] Distributed zeroth-order optimization: Convergence rates that match centralized counterpart
    Yuan, Deming
    Wang, Lei
    Proutiere, Alexandre
    Shi, Guodong
    AUTOMATICA, 2024, 159
  • [50] Communication-Efficient Stochastic Zeroth-Order Optimization for Federated Learning
    Fang, Wenzhi
    Yu, Ziyi
    Jiang, Yuning
    Shi, Yuanming
    Jones, Colin N.
    Zhou, Yong
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2022, 70 : 5058 - 5073