Gradient-Free Methods for Deterministic and Stochastic Nonsmooth Nonconvex Optimization

被引:0
|
作者
Lin, Tianyi [1 ]
Zheng, Zeyu [1 ]
Jordan, Michael I. [1 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA 94720 USA
关键词
CONVEX-OPTIMIZATION; SUBGRADIENT METHODS; SAMPLING ALGORITHM; ZEROTH-ORDER; CONVERGENCE; COMPOSITE;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Nonsmooth nonconvex optimization problems broadly emerge in machine learning and business decision making, whereas two core challenges impede the development of efficient solution methods with finite-time convergence guarantee: the lack of computationally tractable optimality criterion and the lack of computationally powerful oracles. The contributions of this paper are two-fold. First, we establish the relationship between the celebrated Goldstein subdifferential [46] and uniform smoothing, thereby providing the basis and intuition for the design of gradient-free methods that guarantee the finite-time convergence to a set of Goldstein stationary points. Second, we propose the gradient-free method (GFM) and stochastic GFM for solving a class of nonsmooth nonconvex optimization problems and prove that both of them can return a (delta, epsilon)-Goldstein stationary point of a Lipschitz function f at an expected convergence rate at O(d(3/2)delta(-1)epsilon(-4)) where d is the problem dimension. Two-phase versions of GFM and SGFM are also proposed and proven to achieve improved large-deviation results. Finally, we demonstrate the effectiveness of 2-SGFM on training ReLU neural networks with the MINST dataset.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Stochastic subgradient algorithm for nonsmooth nonconvex optimization
    Yalcin, Gulcin Dinc
    [J]. JOURNAL OF APPLIED MATHEMATICS AND COMPUTING, 2024, 70 (01) : 317 - 334
  • [42] Distributed gradient-free and projection-free algorithm for stochastic constrained optimization
    Hou J.
    Zeng X.
    Chen C.
    [J]. Autonomous Intelligent Systems, 2024, 4 (01):
  • [43] Level set methods for gradient-free optimization of metasurface arrays
    Alex Saad-Falcon
    Christopher Howard
    Justin Romberg
    Kenneth Allen
    [J]. Scientific Reports, 14 (1)
  • [44] A unified analysis of stochastic gradient-free Frank-Wolfe methods
    Guo, Jiahong
    Liu, Huiling
    Xiao, Xiantao
    [J]. INTERNATIONAL TRANSACTIONS IN OPERATIONAL RESEARCH, 2022, 29 (01) : 63 - 86
  • [45] Learning Supervised PageRank with Gradient-Based and Gradient-Free Optimization Methods
    Bogolubsky, Lev
    Gusev, Gleb
    Raigorodskii, Andrei
    Tikhonov, Aleksey
    Zhukovskii, Maksim
    Dvurechensky, Pavel
    Gasnikov, Alexander
    Nesterov, Yurii
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [46] Convergence of the gradient sampling algorithm for nonsmooth nonconvex optimization
    Kiwiel, Krzysztof C.
    [J]. SIAM JOURNAL ON OPTIMIZATION, 2007, 18 (02) : 379 - 388
  • [47] Gradient set splitting in nonconvex nonsmooth numerical optimization
    Gaudioso, Manlio
    Gorgone, Enrico
    [J]. OPTIMIZATION METHODS & SOFTWARE, 2010, 25 (01): : 59 - 74
  • [48] A robust gradient sampling algorithm for nonsmooth, nonconvex optimization
    Burke, JV
    Lewis, AS
    Overton, ML
    [J]. SIAM JOURNAL ON OPTIMIZATION, 2005, 15 (03) : 751 - 779
  • [49] Stability and Generalization of Stochastic Optimization with Nonconvex and Nonsmooth Problems
    Lei, Yunwen
    [J]. THIRTY SIXTH ANNUAL CONFERENCE ON LEARNING THEORY, VOL 195, 2023, 195 : 191 - 227
  • [50] A STOCHASTIC SEMISMOOTH NEWTON METHOD FOR NONSMOOTH NONCONVEX OPTIMIZATION
    Milzarek, Andre
    Xiao, Xiantao
    Cen, Shicong
    Wen, Zaiwen
    Ulbrich, Michael
    [J]. SIAM JOURNAL ON OPTIMIZATION, 2019, 29 (04) : 2916 - 2948