Adaptive learning rate optimization algorithms with dynamic bound based on Barzilai-Borwein method

被引:8
|
作者
Wang, Zhi-Jun [1 ]
Gao, He-Bei [2 ]
Wang, Xiang-Hong [3 ]
Zhao, Shuai-Ye [1 ]
Li, Hong [4 ]
Zhang, Xiao-Qin [4 ]
机构
[1] Wenzhou Univ, Coll Comp Sci & Artificial Intelligence, Wenzhou 325035, Peoples R China
[2] Wenzhou Polytech, Coll Artificial Intelligence, Wenzhou 325035, Peoples R China
[3] Shanghai Tech Inst Elect & Informat, Shanghai 201411, Peoples R China
[4] Wenzhou Univ, Key Lab Intelligent Informat Safety & Emergency Zh, Wenzhou 325035, Peoples R China
基金
中国国家自然科学基金;
关键词
Barzilai-Borwein; Adaptive gradient method; Dynamic bound of learning rate; Deep learning;
D O I
10.1016/j.ins.2023.03.050
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The training effect of the neural network model is directly influenced by optimization algorithms. The Barzilai-Borwein(BB) method is used in the stochastic gradient descent (SGD) and other deep learning optimization algorithms because of its outstanding performance in terms of convergence speed. In order to improve the stability of BB step size and the output quality of network model in deep learning, this paper presents two improved optimization algorithms based on the BB method: BBbound and AdaBBbound. BBbound reduces the floating BB step size by generating the upper bound of the step in the current iteration. It avoids the occasional occurrence of long steps; AdaBBbound is an adaptive gradient method based on BB method with dynamic bound. The algorithm has a fast convergence effect in the early stage, then calculates learning rate and smoothly transitions to SGD by setting the better conditions of the BB method. We analyzed the performance of the two algorithms based on the initial conditions and the learning rate change curve. Meanwhile, we tested our algorithms on popular network models such as ResNet and DenseNet. The results showed that the new optimization algorithms have achieved high stability and significantly improved the performance of the network model.
引用
收藏
页码:42 / 54
页数:13
相关论文
共 50 条
  • [1] Adaptive learning rate algorithms based on the improved Barzilai–Borwein method
    Wang, Zhi-Jun
    Li, Hong
    Xu, Zhou-Xiang
    Zhao, Shuai-Ye
    Wang, Peng-Jun
    Gao, He-Bei
    [J]. Pattern Recognition, 2025, 160
  • [2] On the Barzilai-Borwein method
    Fletcher, R
    [J]. OPTIMIZATION AND CONTROL WITH APPLICATIONS, 2005, 96 : 235 - 256
  • [3] An adaptive nonmonotone global Barzilai-Borwein gradient method for unconstrained optimization
    Nosratipour, Hadi
    Fard, Omid Solaymani
    Borzabadi, Akbar Hashemi
    [J]. OPTIMIZATION, 2017, 66 (04) : 641 - 655
  • [4] A NONMONOTONE FILTER BARZILAI-BORWEIN METHOD FOR OPTIMIZATION
    Zhang, Yan
    Sun, Wenyu
    Qi, Liqun
    [J]. ASIA-PACIFIC JOURNAL OF OPERATIONAL RESEARCH, 2010, 27 (01) : 55 - 69
  • [5] The cyclic Barzilai-Borwein method for unconstrained optimization
    Dai, Yu-Hong
    Hager, William W.
    Schittkowski, Klaus
    Zhang, Hongchao
    [J]. IMA JOURNAL OF NUMERICAL ANALYSIS, 2006, 26 (03) : 604 - 627
  • [6] A Barzilai-Borwein descent method for multiobjective optimization problems
    Chen, Jian
    Tang, Liping
    Yang, Xinmin
    [J]. EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2023, 311 (01) : 196 - 209
  • [7] On the acceleration of the Barzilai-Borwein method
    Huang, Yakui
    Dai, Yu-Hong
    Liu, Xin-Wei
    Zhang, Hongchao
    [J]. COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2022, 81 (03) : 717 - 740
  • [8] STABILIZED BARZILAI-BORWEIN METHOD
    Burdakov, Oleg
    Dai, Yuhong
    Huang, Na
    [J]. JOURNAL OF COMPUTATIONAL MATHEMATICS, 2019, 37 (06) : 916 - 936
  • [9] A DIRECT SEARCH FRAME-BASED ADAPTIVE BARZILAI-BORWEIN METHOD
    Fang, Xiaowei
    Ni, Qin
    [J]. JOURNAL OF COMPUTATIONAL MATHEMATICS, 2015, 33 (02) : 179 - 190
  • [10] Subspace Barzilai-Borwein Gradient Method for Large-Scale Bound Constrained Optimization
    Yunhai Xiao
    Qingjie Hu
    [J]. Applied Mathematics and Optimization, 2008, 58 : 275 - 290