Finding Second-Order Stationary Points in Nonconvex-Strongly-Concave Minimax Optimization

被引:0
|
作者
Luo, Luo [1 ]
Li, Yujun [2 ]
Chen, Cheng [3 ]
机构
[1] Fudan Univ, Sch Data Sci, Shanghai, Peoples R China
[2] Huawei Technol Co Ltd, Noahs Ark Lab, Shenzhen, Peoples R China
[3] Nanyang Technol Univ, Sch Phys & Math Sci, Singapore, Singapore
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We study the smooth minimax optimization problem minx maxy f(x, y), where f is l-smooth, strongly-concave in y but possibly nonconvex in x. Most of existing works focus on finding the first-order stationary points of the function f(x, y) or its primal function P(x), Delta= max(y) f(x, y), but few of them focus on achieving second-order stationary points. In this paper, we propose a novel approach for minimax optimization, called Minimax Cubic Newton (MCN), which could find an (epsilon, kappa(1.5) root rho epsilon)-second-order stationary point of P(x) with calling O(kappa(1.5)root rho epsilon(-1.5)) times of second-order oracles and (O) over tilde(kappa(2) root rho epsilon(-1.5)) times of first-order oracles, where kappa is the condition number and rho is the Lipschitz continuous constant for the Hessian of f(x, y). In addition, we propose an inexact variant of MCN for high-dimensional problems to avoid calling expensive second-order oracles. Instead, our method solves the cubic sub-problem inexactly via gradient descent and matrix Chebyshev expansion. This strategy still obtains the desired approximate second-order stationary point with high probability but only requires (O) over tilde(kappa(1.5) l epsilon(-2)) Hessian-vector oracle calls and (O) over tilde(kappa(2)root rho epsilon(-1.5)) first-order oracle calls. To the best of our knowledge, this is the first work that considers the non-asymptotic convergence behavior of finding second-order stationary points for minimax problems without the convex-concave assumptions.
引用
收藏
页数:13
相关论文
共 50 条
  • [21] Second-Order Guarantees of Stochastic Gradient Descent in Nonconvex Optimization
    Vlaski, Stefan
    Sayed, Ali H.
    IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2022, 67 (12) : 6489 - 6504
  • [22] Inexact proximal stochastic second-order methods for nonconvex composite optimization
    Wang, Xiao
    Zhang, Hongchao
    OPTIMIZATION METHODS & SOFTWARE, 2020, 35 (04): : 808 - 835
  • [23] An adaptive high order method for finding third-order critical points of nonconvex optimization
    Xihua Zhu
    Jiangze Han
    Bo Jiang
    Journal of Global Optimization, 2022, 84 : 369 - 392
  • [24] An adaptive high order method for finding third-order critical points of nonconvex optimization
    Zhu, Xihua
    Han, Jiangze
    Jiang, Bo
    JOURNAL OF GLOBAL OPTIMIZATION, 2022, 84 (02) : 369 - 392
  • [25] A subclass of generating set search with convergence to second-order stationary points
    Abramson, Mark A.
    Frimannslund, Lennart
    Steihaug, Trond
    OPTIMIZATION METHODS & SOFTWARE, 2014, 29 (05): : 900 - 918
  • [26] Augmented Lagrangians with constrained subproblems and convergence to second-order stationary points
    E. G. Birgin
    G. Haeser
    A. Ramos
    Computational Optimization and Applications, 2018, 69 : 51 - 75
  • [27] Convergence to second order stationary points in inequality constrained optimization
    Facchinei, F
    Lucidi, S
    MATHEMATICS OF OPERATIONS RESEARCH, 1998, 23 (03) : 746 - 766
  • [28] Augmented Lagrangians with constrained subproblems and convergence to second-order stationary points
    Birgin, E. G.
    Haeser, G.
    Ramos, A.
    COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2018, 69 (01) : 51 - 75
  • [29] Convergence of mesh adaptive direct search to second-order stationary points
    Abramson, Mark A.
    Audet, Charles
    SIAM JOURNAL ON OPTIMIZATION, 2006, 17 (02) : 606 - 619
  • [30] Private (Stochastic) Non-Convex Optimization Revisited: Second-Order Stationary Points and Excess Risks
    Ganesh, Arun
    Liu, Daogao
    Oh, Sewoong
    Thakurta, Abhradeep
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,