Difference-enhanced adaptive momentum methods for non-convex stochastic optimization in image classification

被引:0
|
作者
Ouyang, Chen [1 ,2 ]
Jian, Ailun [3 ]
Zhao, Xiong [1 ,2 ]
Yuan, Gonglin [1 ,2 ]
机构
[1] Guangxi Univ, Sch Math & Informat Sci, Nanning 530004, Guangxi, Peoples R China
[2] Guangxi Univ, Ctr Appl Math Guangxi Guangxi Univ, Nanning 530004, Guangxi, Peoples R China
[3] Hangzhou Dianzi Univ, Sch Sci, Hangzhou 310027, Zhejiang, Peoples R China
关键词
Adaptive momentum methods; Non-convex; Deep learning; Image classification; CONJUGATE-GRADIENT METHOD; ALGORITHMS; DESCENT;
D O I
10.1016/j.dsp.2025.105118
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Stochastic gradient descent with momentum (SGDM) is a classic optimization method that determines the update direction using a moving average of the gradient over historical steps. However, SGDM suffers from slow convergence. In 2022, Yuan et al. [6] proposed stochastic gradient descent with momentum and difference (SGDMD), which incorporates the concept of differences to adjust the convergence direction and accelerate the optimization process. Despite its improvements, SGDMD requires careful parameter tuning and is prone to oscillations due to the difference mechanism. In this work, we introduce a new momentum method: stochastic gradient descent with adaptive momentum and difference (SGDAMD). Compared to SGDMD, SGDAMD demonstrates superior performance in experiments, achieving greater stability in terms of both loss values and accuracy in deep learning image classification tasks. Additionally, SGDAMD attains a sublinear convergence rate in non-convex settings while requiring less restrictive assumptions than standard smoothness conditions. These features underscore the algorithm's efficiency and effectiveness in addressing complex optimization challenges.
引用
收藏
页数:9
相关论文
共 50 条
  • [41] Tensor recovery based on rank adaptive and non-convex methods
    Liu C.
    Zhang H.
    Fan H.
    Li Y.
    Optik, 2023, 292
  • [42] Image reconstruction from few views by non-convex optimization
    Sidky, Emil Y.
    Chartrand, Rick
    Pan, Xiaochuan
    2007 IEEE NUCLEAR SCIENCE SYMPOSIUM CONFERENCE RECORD, VOLS 1-11, 2007, : 3526 - +
  • [43] Hyperspectral fruit image restoration using non-convex optimization
    Shanthini, K. S.
    George, Sudhish N.
    George, Sony
    Devassy, Binu Melit
    JOURNAL OF FOOD ENGINEERING, 2023, 358
  • [44] A new non-convex sparse optimization method for image restoration
    Peng Wu
    Dequan Li
    Signal, Image and Video Processing, 2023, 17 : 3829 - 3836
  • [45] A new non-convex sparse optimization method for image restoration
    Wu, Peng
    Li, Dequan
    SIGNAL IMAGE AND VIDEO PROCESSING, 2023, 17 (07) : 3829 - 3836
  • [46] A Momentum-Based Adaptive Primal-Dual Stochastic Gradient Method for Non-Convex Programs with Expectation Constraints
    Qi, Rulei
    Xue, Dan
    Zhai, Yujia
    MATHEMATICS, 2024, 12 (15)
  • [47] Improved Variance Reduction Methods for Riemannian Non-Convex Optimization
    Han, Andi
    Gao, Junbin
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (11) : 7610 - 7623
  • [49] Stochastic Optimization for Learning Non-convex Linear Support Vector Machines
    Chen, Jifei
    Wang, Jiabao
    Zhang, Yafei
    Lu, Jianjiang
    Li, Yang
    2012 IEEE INTERNATIONAL CONFERENCE ON INTELLIGENT CONTROL, AUTOMATIC DETECTION AND HIGH-END EQUIPMENT (ICADE), 2012, : 35 - 39
  • [50] Surrogate Losses for Online Learning of Stepsizes in Stochastic Non-Convex Optimization
    Zhuang, Zhenxun
    Cutkosky, Ashok
    Orabona, Francesco
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 97, 2019, 97