Difference-enhanced adaptive momentum methods for non-convex stochastic optimization in image classification

被引:0
|
作者
Ouyang, Chen [1 ,2 ]
Jian, Ailun [3 ]
Zhao, Xiong [1 ,2 ]
Yuan, Gonglin [1 ,2 ]
机构
[1] Guangxi Univ, Sch Math & Informat Sci, Nanning 530004, Guangxi, Peoples R China
[2] Guangxi Univ, Ctr Appl Math Guangxi Guangxi Univ, Nanning 530004, Guangxi, Peoples R China
[3] Hangzhou Dianzi Univ, Sch Sci, Hangzhou 310027, Zhejiang, Peoples R China
关键词
Adaptive momentum methods; Non-convex; Deep learning; Image classification; CONJUGATE-GRADIENT METHOD; ALGORITHMS; DESCENT;
D O I
10.1016/j.dsp.2025.105118
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Stochastic gradient descent with momentum (SGDM) is a classic optimization method that determines the update direction using a moving average of the gradient over historical steps. However, SGDM suffers from slow convergence. In 2022, Yuan et al. [6] proposed stochastic gradient descent with momentum and difference (SGDMD), which incorporates the concept of differences to adjust the convergence direction and accelerate the optimization process. Despite its improvements, SGDMD requires careful parameter tuning and is prone to oscillations due to the difference mechanism. In this work, we introduce a new momentum method: stochastic gradient descent with adaptive momentum and difference (SGDAMD). Compared to SGDMD, SGDAMD demonstrates superior performance in experiments, achieving greater stability in terms of both loss values and accuracy in deep learning image classification tasks. Additionally, SGDAMD attains a sublinear convergence rate in non-convex settings while requiring less restrictive assumptions than standard smoothness conditions. These features underscore the algorithm's efficiency and effectiveness in addressing complex optimization challenges.
引用
收藏
页数:9
相关论文
共 50 条
  • [31] Asymptotic Study of Stochastic Adaptive Algorithms in Non-convex Landscape
    Gadat, Sébastien
    Gavra, Ioana
    Journal of Machine Learning Research, 2022, 23
  • [32] Asymptotic Study of Stochastic Adaptive Algorithms in Non-convex Landscape
    Gadat, Sebastien
    Gavra, Ioana
    JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23
  • [33] Variance Reduced Methods for Non-Convex Composition Optimization
    Liu, Liu
    Liu, Ji
    Tao, Dacheng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (09) : 5813 - 5825
  • [34] Differentially Private Stochastic Optimization: New Results in Convex and Non-Convex Settings
    Bassily, Raef
    Guzman, Cristobal
    Menart, Michael
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [35] Private Stochastic Non-convex Optimization with Improved Utility Rates
    Zhang, Qiuchen
    Ma, Jing
    Lou, Jian
    Xiong, Li
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 3370 - 3376
  • [36] Adaptive non-convex total variation regularisation for image restoration
    Fu, S.
    Zhang, C.
    ELECTRONICS LETTERS, 2010, 46 (13) : 907 - U56
  • [37] Non-convex nonlocal adaptive tight frame image deblurring
    Shen, Zhengwei
    IET IMAGE PROCESSING, 2022, 16 (07) : 1908 - 1923
  • [38] A new non-adaptive optimization method: Stochastic gradient descent with momentum and difference
    Wei Yuan
    Fei Hu
    Liangfu Lu
    Applied Intelligence, 2022, 52 : 3939 - 3953
  • [39] A new non-adaptive optimization method: Stochastic gradient descent with momentum and difference
    Yuan, Wei
    Hu, Fei
    Lu, Liangfu
    APPLIED INTELLIGENCE, 2022, 52 (04) : 3939 - 3953
  • [40] Adaptive Negative Curvature Descent with Applications in Non-convex Optimization
    Liu, Mingrui
    Li, Zhe
    Wang, Xiaoyu
    Yi, Jinfeng
    Yang, Tianbao
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31