Gradient-Free Methods for Deterministic and Stochastic Nonsmooth Nonconvex Optimization

被引:0
|
作者
Lin, Tianyi [1 ]
Zheng, Zeyu [1 ]
Jordan, Michael I. [1 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA 94720 USA
关键词
CONVEX-OPTIMIZATION; SUBGRADIENT METHODS; SAMPLING ALGORITHM; ZEROTH-ORDER; CONVERGENCE; COMPOSITE;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Nonsmooth nonconvex optimization problems broadly emerge in machine learning and business decision making, whereas two core challenges impede the development of efficient solution methods with finite-time convergence guarantee: the lack of computationally tractable optimality criterion and the lack of computationally powerful oracles. The contributions of this paper are two-fold. First, we establish the relationship between the celebrated Goldstein subdifferential [46] and uniform smoothing, thereby providing the basis and intuition for the design of gradient-free methods that guarantee the finite-time convergence to a set of Goldstein stationary points. Second, we propose the gradient-free method (GFM) and stochastic GFM for solving a class of nonsmooth nonconvex optimization problems and prove that both of them can return a (delta, epsilon)-Goldstein stationary point of a Lipschitz function f at an expected convergence rate at O(d(3/2)delta(-1)epsilon(-4)) where d is the problem dimension. Two-phase versions of GFM and SGFM are also proposed and proven to achieve improved large-deviation results. Finally, we demonstrate the effectiveness of 2-SGFM on training ReLU neural networks with the MINST dataset.
引用
收藏
页数:16
相关论文
共 50 条
  • [11] Gradient-free method for nonsmooth distributed optimization
    Li, Jueyou
    Wu, Changzhi
    Wu, Zhiyou
    Long, Qiang
    [J]. JOURNAL OF GLOBAL OPTIMIZATION, 2015, 61 (02) : 325 - 340
  • [12] Gradient-free method for nonsmooth distributed optimization
    Jueyou Li
    Changzhi Wu
    Zhiyou Wu
    Qiang Long
    [J]. Journal of Global Optimization, 2015, 61 : 325 - 340
  • [13] Stochastic generalized gradient method for nonconvex nonsmooth stochastic optimization
    Yu. M. Ermol'ev
    V. I. Norkin
    [J]. Cybernetics and Systems Analysis, 1998, 34 : 196 - 215
  • [14] Stochastic generalized gradient method for nonconvex nonsmooth stochastic optimization
    Ermol'ev, YM
    Norkin, VI
    [J]. CYBERNETICS AND SYSTEMS ANALYSIS, 1998, 34 (02) : 196 - 215
  • [15] Deterministic Nonsmooth Nonconvex Optimization
    Jordan, Michael I.
    Kornowski, Guy
    Lin, Tianyi
    Shamir, Ohad
    Zampetakis, Manolis
    [J]. THIRTY SIXTH ANNUAL CONFERENCE ON LEARNING THEORY, VOL 195, 2023, 195
  • [16] INCREMENTAL GRADIENT-FREE METHOD FOR NONSMOOTH DISTRIBUTED OPTIMIZATION
    Li, Jueyou
    Li, Guoquan
    Wu, Zhiyou
    Wu, Changzhi
    Wang, Xiangyu
    Lee, Jae-Myung
    Jung, Kwang-Hyo
    [J]. JOURNAL OF INDUSTRIAL AND MANAGEMENT OPTIMIZATION, 2017, 13 (04) : 1841 - 1857
  • [17] Gradient-Free Method for Heavily Constrained Nonconvex Optimization
    Shi, Wanli
    Gao, Hongchang
    Gu, Bin
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [18] VARIABLE METRIC PROXIMAL STOCHASTIC VARIANCE REDUCED GRADIENT METHODS FOR NONCONVEX NONSMOOTH OPTIMIZATION
    Yu, Tengteng
    Liu, Xin-wei
    Dai, Yu-hong
    Sun, J. I. E.
    [J]. JOURNAL OF INDUSTRIAL AND MANAGEMENT OPTIMIZATION, 2022, 18 (04) : 2611 - 2631
  • [19] A Simple Proximal Stochastic Gradient Method for Nonsmooth Nonconvex Optimization
    Li, Zhize
    Li, Jian
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [20] Adaptivity of Stochastic Gradient Methods for Nonconvex Optimization
    Horvath, Samuel
    Lei, Lihua
    Richtarik, Peter
    Jordan, Michael I.
    [J]. SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2022, 4 (02): : 634 - 648