A new spectral conjugate gradient method for large-scale unconstrained optimization

被引:43
|
作者
Jian, Jinbao [1 ,2 ]
Chen, Qian [2 ]
Jiang, Xianzhen [1 ]
Zeng, Youfang [2 ]
Yin, Jianghua [2 ]
机构
[1] Yulin Normal Univ, Guangxi Coll & Univ Key Lab Complex Syst Optimiza, Yulin 537000, Peoples R China
[2] Guangxi Univ, Coll Math & Informat Sci, Nanning 530004, Peoples R China
来源
OPTIMIZATION METHODS & SOFTWARE | 2017年 / 32卷 / 03期
关键词
large-scale unconstrained optimization; spectral conjugate gradient method; global convergence; numerical experiments; 90C25; 90C30; CONVERGENCE CONDITIONS; GLOBAL CONVERGENCE; ALGORITHM; DESCENT; MINIMIZATION; PROPERTY;
D O I
10.1080/10556788.2016.1225213
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
The spectral conjugate gradient methods, with simple construction and nice numerical performance, are a kind of effective methods for solving large-scale unconstrained optimization problems. In this paper, based on quasi-Newton direction and quasi-Newton condition, and motivated by the idea of spectral conjugate gradient method as well as Dai-Kou's selecting technique for conjugate parameter [SIAM J. Optim. 23 (2013), pp. 296-320], a new approach for generating spectral parameters is presented, where a new double-truncating technique, which can ensure both the sufficient descent property of the search directions and the bounded property of the sequence of spectral parameters, is introduced. Then a new associated spectral conjugate gradient method for large-scale unconstrained optimization is proposed. Under either the strong Wolfe line search or the generalized Wolfe line search, the proposed method is always globally convergent. Finally, a large number of comparison numerical experiments on large-scale instances from one thousand to two million variables are reported. The numerical results show that the proposed method is more promising.
引用
收藏
页码:503 / 515
页数:13
相关论文
共 50 条