ZEROTH-ORDER REGULARIZED OPTIMIZATION (ZORO): APPROXIMATELY SPARSE GRADIENTS AND ADAPTIVE SAMPLING

被引:7
|
作者
Cai, HanQin [1 ]
McKenzie, Daniel [1 ]
Yin, Wotao [2 ]
Zhang, Zhenliang [3 ]
机构
[1] Univ Calif Los Angeles, Dept Math, Los Angeles, CA 90095 USA
[2] Alibaba US, Damo Acad, Bellevue, WA 98004 USA
[3] Xmotors AI, Mountain View, CA 94043 USA
关键词
zeroth-order optimization; black-box optimization; derivative-free optimization; regularized optimization; sparse gradients; sparse adversarial attack; CONVERGENCE; PARAMETER; SEARCH;
D O I
10.1137/21M1392966
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
We consider the problem of minimizing a high-dimensional objective function, which may include a regularization term, using only (possibly noisy) evaluations of the function. Such optimization is also called derivative-free, zeroth-order, or black-box optimization. We propose a new zeroth-order regularized optimization method, dubbed ZORO. When the underlying gradient is approximately sparse at an iterate, ZORO needs very few objective function evaluations to obtain a new iterate that decreases the objective function. We achieve this with an adaptive, randomized gradient estimator, followed by an inexact proximal-gradient scheme. Under a novel approximately sparse gradient assumption and various different convex settings, we show that the (theoretical and empirical) convergence rate of ZORO is only logarithmically dependent on the problem dimension. Numerical experiments show that ZORO outperforms existing methods with similar assumptions, on both synthetic and real datasets.
引用
收藏
页码:687 / 714
页数:28
相关论文
共 50 条
  • [31] ZEROTH-ORDER STOCHASTIC PROJECTED GRADIENT DESCENT FOR NONCONVEX OPTIMIZATION
    Liu, Sijia
    Li, Xingguo
    Chen, Pin-Yu
    Haupt, Jarvis
    Amini, Lisa
    2018 IEEE GLOBAL CONFERENCE ON SIGNAL AND INFORMATION PROCESSING (GLOBALSIP 2018), 2018, : 1179 - 1183
  • [32] Zeroth-Order Methods for Nondifferentiable, Nonconvex, and Hierarchical Federated Optimization
    Qiu, Yuyang
    Shanbhag, Uday V.
    Yousefian, Farzad
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [33] A Generic Approach for Accelerating Stochastic Zeroth-Order Convex Optimization
    Yu, Xiaotian
    King, Irwin
    Lyu, Michael R.
    Yang, Tianbao
    PROCEEDINGS OF THE TWENTY-SEVENTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2018, : 3040 - 3046
  • [34] A zeroth-order algorithm for distributed optimization with stochastic stripe observations
    Wang, Yinghui
    Zeng, Xianlin
    Zhao, Wenxiao
    Hong, Yiguang
    SCIENCE CHINA-INFORMATION SCIENCES, 2023, 66 (09)
  • [35] Zeroth-Order Negative Curvature Finding: Escaping Saddle Points without Gradients
    Zhang, Hualin
    Xiong, Huan
    Gu, Bin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [36] A zeroth-order algorithm for distributed optimization with stochastic stripe observations
    Yinghui WANG
    Xianlin ZENG
    Wenxiao ZHAO
    Yiguang HONG
    Science China(Information Sciences), 2023, 66 (09) : 297 - 298
  • [37] On the Convergence of Prior-Guided Zeroth-Order Optimization Algorithms
    Cheng, Shuyu
    Wu, Guoqiang
    Zhu, Jun
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [38] First and Zeroth-Order Implementations of the Regularized Newton Method with Lazy Approximated HessiansFirst and zeroth-order regularized Newton method with lazy Hessians...N. Doikov, G. N. Grapiglia
    Nikita Doikov
    Geovani Nunes Grapiglia
    Journal of Scientific Computing, 2025, 103 (1)
  • [39] Automatic controller tuning using a zeroth-order optimization algorithm
    Zalkind, Daniel S.
    Dall'Anese, Emiliano
    Pao, Lucy Y.
    WIND ENERGY SCIENCE, 2020, 5 (04) : 1579 - 1600
  • [40] Zeroth-Order Optimization for Varactor-Tuned Matching Network
    Pirrone, Michelle
    Dall'Anese, Emiliano
    Barton, Taylor
    2022 IEEE/MTT-S INTERNATIONAL MICROWAVE SYMPOSIUM (IMS 2022), 2022, : 502 - 505