A globally convergent proximal Newton-type method in nonsmooth convex optimization

被引:14
|
作者
Mordukhovich, Boris S. [1 ]
Yuan, Xiaoming [2 ]
Zeng, Shangzhi [3 ]
Zhang, Jin [4 ]
机构
[1] Wayne State Univ, Dept Math, Detroit, MI 48202 USA
[2] Univ Hong Kong, Dept Math, Hong Kong, Peoples R China
[3] Univ Victoria, Dept Math & Stat, Victoria, BC, Canada
[4] Southern Univ Sci & Technol, Natl Ctr Appl Math Shenzhen, Dept Math, Shenzhen 518055, Peoples R China
基金
澳大利亚研究理事会; 美国国家科学基金会;
关键词
Nonsmooth convex optimization; Machine learning; Proximal Newton methods; Global and local convergence; Metric subregularity; METRIC SUBREGULARITY; GENERALIZED EQUATIONS; REGULARITY; ALGORITHM;
D O I
10.1007/s10107-022-01797-5
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
The paper proposes and justifies a new algorithm of the proximal Newton type to solve a broad class of nonsmooth composite convex optimization problems without strong convexity assumptions. Based on advanced notions and techniques of variational analysis, we establish implementable results on the global convergence of the proposed algorithm as well as its local convergence with superlinear and quadratic rates. For certain structured problems, the obtained local convergence conditions do not require the local Lipschitz continuity of the corresponding Hessian mappings that is a crucial assumption used in the literature to ensure a superlinear convergence of other algorithms of the proximal Newton type. The conducted numerical experiments of solving the l(1) regularized logistic regression model illustrate the possibility of applying the proposed algorithm to deal with practically important problems.
引用
收藏
页码:899 / 936
页数:38
相关论文
共 50 条