Research on three-step accelerated gradient algorithm in deep learning

被引:0
|
作者
Lian, Yongqiang [1 ,2 ]
Tang, Yincai [1 ]
Zhou, Shirong [1 ]
机构
[1] East China Normal Univ, Sch Stat, KLATASDS MOE, Shanghai, Peoples R China
[2] East China Normal Univ, Sch Stat, KLATASDS MOE, Shanghai 200062, Peoples R China
基金
中国国家自然科学基金;
关键词
Accelerated algorithm; backpropagation; deep learning; learning rate; momentum; stochastic gradient descent;
D O I
10.1080/24754269.2020.1846414
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
Gradient descent (GD) algorithm is the widely used optimisation method in training machine learning and deep learning models. In this paper, based on GD, Polyak's momentum (PM), and Nesterov accelerated gradient (NAG), we give the convergence of the algorithms from an initial value to the optimal value of an objective function in simple quadratic form. Based on the convergence property of the quadratic function, two sister sequences of NAG's iteration and parallel tangent methods in neural networks, the three-step accelerated gradient (TAG) algorithm is proposed, which has three sequences other than two sister sequences. To illustrate the performance of this algorithm, we compare the proposed algorithm with the three other algorithms in quadratic function, high-dimensional quadratic functions, and nonquadratic function. Then we consider to combine the TAG algorithm to the backpropagation algorithm and the stochastic gradient descent algorithm in deep learning. For conveniently facilitate the proposed algorithms, we rewite the R package 'neuralnet' and extend it to 'supneuralnet'. All kinds of deep learning algorithms in this paper are included in 'supneuralnet' package. Finally, we show our algorithms are superior to other algorithms in four case studies.
引用
收藏
页码:40 / 57
页数:18
相关论文
共 50 条
  • [1] On the Convergence Rate of Three-Step Gradient Methods
    Liu, Ziyi
    Chai, Li
    Liu, Bing
    [J]. 2020 EUROPEAN CONTROL CONFERENCE (ECC 2020), 2020, : 1789 - 1794
  • [2] A hierarchical three-step superpixels and deep learning framework for skin lesion classification
    Afza, Farhat
    Sharif, Muhammad
    Mittal, Mamta
    Khan, Muhammad Attique
    Hemanth, Jude
    [J]. METHODS, 2022, 202 : 88 - 102
  • [3] A novel three-step deep learning approach for the classification of breast cancer histopathological images
    Kolla, Bhavannarayanna
    Venugopal, P.
    [J]. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2023, 45 (06) : 10477 - 10495
  • [4] Recognition and counting of wheat mites in wheat fields by a three-step deep learning method
    Chen, Peng
    Li, WeiLu
    Yao, SiJie
    Ma, Chun
    Zhang, Jun
    Wang, Bing
    Zheng, ChunHou
    Xie, ChengJun
    Liang, Dong
    [J]. NEUROCOMPUTING, 2021, 437 : 21 - 30
  • [5] Fast three-step phase-shifting algorithm
    Huang, Peisen S.
    Zhang, Song
    [J]. APPLIED OPTICS, 2006, 45 (21) : 5086 - 5091
  • [6] Double three-step phase-shifting algorithm
    Huang, PS
    Hu, QYJ
    Chiang, FP
    [J]. APPLIED OPTICS, 2002, 41 (22) : 4503 - 4509
  • [7] An Effective Three-step Search Algorithm for Motion Estimation
    Sun Ning-ning
    Fan Chao
    Xia Xu
    [J]. 2009 IEEE INTERNATIONAL SYMPOSIUM ON IT IN MEDICINE & EDUCATION, VOLS 1 AND 2, PROCEEDINGS, 2009, : 400 - 403
  • [8] Machine learning accelerated holographic near-eye display system based on three-step diffraction
    Shen, Chentianfei
    Shen, Tong
    Chen, Qi
    Zheng, Jihong
    [J]. AOPC 2021: MICRO-OPTICS AND MOEMS, 2021, 12066
  • [9] A three-step synthetic extraction algorithm for transmission lines
    Zhang, Dongsheng
    Wang, Jing
    Han, Rongjuan
    Tian, Aifen
    [J]. PROCEEDINGS OF THE INSTITUTION OF MECHANICAL ENGINEERS PART C-JOURNAL OF MECHANICAL ENGINEERING SCIENCE, 2019, 233 (17) : 6218 - 6228
  • [10] Equivalence in international business research: A three-step approach
    Angelo M. Solarino
    Peter J. Buckley
    [J]. Journal of International Business Studies, 2023, 54 : 550 - 567