Subsampled inexact Newton methods for minimizing large sums of convex functions

被引:12
|
作者
Bellavia, Stefania [1 ]
Krejic, Natasa [2 ]
Jerinkic, Natasa Krklec [2 ]
机构
[1] Univ Florence, Dept Ind Engn, Viale Morgagni 40-44, I-50134 Florence, Italy
[2] Univ Novi Sad, Fac Sci, Dept Math & Informat, Trg Dositeja Obradov 4, Novi Sad 21000, Serbia
关键词
inexact Newton; subsampled Hessian; superlinear convergence; global convergence; mean square convergence; LINE-SEARCH; OPTIMIZATION METHODS; SAMPLE-SIZE; CONVERGENCE; ALGORITHM;
D O I
10.1093/imanum/drz027
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
This paper deals with the minimization of a large sum of convex functions by inexact Newton (IN) methods employing subsampled functions, gradients and Hessian approximations. The conjugate gradient method is used to compute the IN step and global convergence is enforced by a nonmonotone line-search procedure. The aim is to obtain methods with affordable costs and fast convergence. Assuming strongly convex functions, R-linear convergence and worst-case iteration complexity of the procedure are investigated when functions and gradients are approximated with increasing accuracy. A set of rules for the forcing parameters and subsample Hessian sizes are derived that ensure local q-linear/q-superlinear convergence of the proposed method. The random choice of the Hessian subsample is also considered and convergence in the mean square, both for finite and infinite sums of functions, is proved. Finally, the analysis of global convergence with asymptotic R-linear rate is extended to the case of the sum of convex functions and strongly convex objective function. Numerical results on well-known binary classification problems are also given. Adaptive strategies for selecting forcing terms and Hessian subsample size, streaming out of the theoretical analysis, are employed and the numerical results show that they yield effective IN methods.
引用
收藏
页码:2309 / 2341
页数:33
相关论文
共 50 条