Gradient-Sensitive Optimization for Convolutional Neural Networks

被引:8
|
作者
Liu, Zhipeng [1 ]
Feng, Rui [1 ]
Li, Xiuhan [1 ]
Wang, Wei [1 ]
Wu, Xiaoling [1 ]
机构
[1] Nanjing Med Univ, Sch Biomed Engn & Informat, Key Lab Clin Engn, Nanjing 211166, Peoples R China
关键词
All Open Access; Gold;
D O I
10.1155/2021/6671830
中图分类号
Q [生物科学];
学科分类号
07 ; 0710 ; 09 ;
摘要
Convolutional neural networks (CNNs) are effective models for image classification and recognition. Gradient descent optimization (GD) is the basic algorithm for CNN model optimization. Since GD appeared, a series of improved algorithms have been derived. Among these algorithms, adaptive moment estimation (Adam) has been widely recognized. However, local changes are ignored in Adam to some extent. In this paper, we introduce an adaptive learning rate factor based on current and recent gradients. According to this factor, we can dynamically adjust the learning rate of each independent parameter to adaptively adjust the global convergence process. We use the factor to adjust the learning rate for each parameter. The convergence of the proposed algorithm is proven by using the regret bound approach of the online learning framework. In the experimental section, comparisons are conducted between the proposed algorithm and other existing algorithms, such as AdaGrad, RMSprop, Adam, diffGrad, and AdaHMG, on test functions and the MNIST dataset. The results show that Adam and RMSprop combined with our algorithm can not only find the global minimum faster in the experiment using the test function but also have a better convergence curve and higher test set accuracy in experiments using datasets. Our algorithm is a supplement to the existing gradient descent algorithms, which can be combined with many other existing gradient descent algorithms to improve the efficiency of iteration, speed up the convergence of the cost function, and improve the final recognition rate.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] SURFACE BEHAVIOR OF GRADIENT-SENSITIVE LIQUIDS
    BLINOWSKI, A
    [J]. ARCHIVES OF MECHANICS, 1973, 25 (02): : 259 - 268
  • [2] ON THE DISTINCTION BETWEEN VELOCITY-SENSITIVE AND PRESSURE GRADIENT-SENSITIVE MICROPHONES
    SCHULTZ, TJ
    [J]. JOURNAL OF THE ACOUSTICAL SOCIETY OF AMERICA, 1956, 28 (03): : 498 - 498
  • [3] Applying Gradient Descent in Convolutional Neural Networks
    Cui, Nan
    [J]. 2ND INTERNATIONAL CONFERENCE ON MACHINE VISION AND INFORMATION TECHNOLOGY (CMVIT 2018), 2018, 1004
  • [4] A Comparative Analysis of Gradient Descent-Based Optimization Algorithms on Convolutional Neural Networks
    Dogo, E. M.
    Afolabi, O. J.
    Nwulu, N. I.
    Twala, B.
    Aigbavboa, C. O.
    [J]. PROCEEDINGS OF THE 2018 INTERNATIONAL CONFERENCE ON COMPUTATIONAL TECHNIQUES, ELECTRONICS AND MECHANICAL SYSTEMS (CTEMS), 2018, : 92 - 99
  • [5] Abs-CAM: a gradient optimization interpretable approach for explanation of convolutional neural networks
    Chunyan Zeng
    Kang Yan
    Zhifeng Wang
    Yan Yu
    Shiyan Xia
    Nan Zhao
    [J]. Signal, Image and Video Processing, 2023, 17 : 1069 - 1076
  • [6] Using Particle Swarm Optimization with Gradient Descent for Parameter Learning in Convolutional Neural Networks
    Wessels, Steven
    van der Haar, Dustin
    [J]. PROGRESS IN PATTERN RECOGNITION, IMAGE ANALYSIS, COMPUTER VISION, AND APPLICATIONS, CIARP 2021, 2021, 12702 : 119 - 128
  • [7] Abs-CAM: a gradient optimization interpretable approach for explanation of convolutional neural networks
    Zeng, Chunyan
    Yan, Kang
    Wang, Zhifeng
    Yu, Yan
    Xia, Shiyan
    Zhao, Nan
    [J]. SIGNAL IMAGE AND VIDEO PROCESSING, 2023, 17 (04) : 1069 - 1076
  • [8] Convolutional neural networks with fractional order gradient method
    Sheng, Dian
    Wei, Yiheng
    Chen, Yuquan
    Wang, Yong
    [J]. NEUROCOMPUTING, 2020, 408 : 42 - 50
  • [9] GRADIENT LOCAL BINARY PATTERN FOR CONVOLUTIONAL NEURAL NETWORKS
    Tang, Jialiang
    Jiang, Ning
    Yu, Wenxin
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 744 - 748
  • [10] Calibrated Stochastic Gradient Descent for Convolutional Neural Networks
    Zhuo, Li'an
    Zhang, Baochang
    Chen, Chen
    Ye, Qixiang
    Liu, Jianzhuang
    Doermann, David
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 9348 - 9355