Alternating and Parallel Proximal Gradient Methods for Nonsmooth, Nonconvex Minimax: A Unified Convergence Analysis

被引:1
|
作者
Cohen, Eyal [1 ]
Teboulle, Marc [1 ]
机构
[1] Tel Aviv Univ, Sch Math Sci, IL-69978 Tel Aviv, Israel
基金
以色列科学基金会;
关键词
nonconvex nonsmooth minimax; nonsmooth minimization-maximization; proximal gradient method; Kurdyka-Lojasiewicz property; Bregman distance; global convergence; convergence rate; 1ST-ORDER METHODS; MIN; ALGORITHM; MINIMIZATION; CONTINUITY;
D O I
10.1287/moor.2022.0294
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
There is growing interest in nonconvex minimax problems that is driven by an abundance of applications. Our focus is on nonsmooth, nonconvex-strongly concave minimax, thus departing from the more common weakly convex and smooth models assumed in the recent literature. We present proximal gradient schemes with either parallel or alternating steps. We show that both methods can be analyzed through a single scheme within a unified analysis that relies on expanding a general convergence mechanism used for analyzing nonconvex, nonsmooth optimization problems. In contrast to the current literature, which focuses on the complexity of obtaining nearly approximate stationary solutions, we prove subsequence convergence to a critical point of the primal objective and global convergence when the latter is semialgebraic. Furthermore, the complexity results we provide are with respect to approximate stationary solutions. Lastly, we expand the scope of problems that can be addressed by generalizing one of the steps with a Bregman proximal gradient update, and together with a few adjustments to the analysis, this allows us to extend the convergence and complexity results to this broader setting.
引用
收藏
页码:1 / 28
页数:29
相关论文
共 50 条