Difference-enhanced adaptive momentum methods for non-convex stochastic optimization in image classification

被引:0
|
作者
Ouyang, Chen [1 ,2 ]
Jian, Ailun [3 ]
Zhao, Xiong [1 ,2 ]
Yuan, Gonglin [1 ,2 ]
机构
[1] Guangxi Univ, Sch Math & Informat Sci, Nanning 530004, Guangxi, Peoples R China
[2] Guangxi Univ, Ctr Appl Math Guangxi Guangxi Univ, Nanning 530004, Guangxi, Peoples R China
[3] Hangzhou Dianzi Univ, Sch Sci, Hangzhou 310027, Zhejiang, Peoples R China
关键词
Adaptive momentum methods; Non-convex; Deep learning; Image classification; CONJUGATE-GRADIENT METHOD; ALGORITHMS; DESCENT;
D O I
10.1016/j.dsp.2025.105118
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Stochastic gradient descent with momentum (SGDM) is a classic optimization method that determines the update direction using a moving average of the gradient over historical steps. However, SGDM suffers from slow convergence. In 2022, Yuan et al. [6] proposed stochastic gradient descent with momentum and difference (SGDMD), which incorporates the concept of differences to adjust the convergence direction and accelerate the optimization process. Despite its improvements, SGDMD requires careful parameter tuning and is prone to oscillations due to the difference mechanism. In this work, we introduce a new momentum method: stochastic gradient descent with adaptive momentum and difference (SGDAMD). Compared to SGDMD, SGDAMD demonstrates superior performance in experiments, achieving greater stability in terms of both loss values and accuracy in deep learning image classification tasks. Additionally, SGDAMD attains a sublinear convergence rate in non-convex settings while requiring less restrictive assumptions than standard smoothness conditions. These features underscore the algorithm's efficiency and effectiveness in addressing complex optimization challenges.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Distributed stochastic gradient tracking methods with momentum acceleration for non-convex optimization
    Juan Gao
    Xin-Wei Liu
    Yu-Hong Dai
    Yakui Huang
    Junhua Gu
    Computational Optimization and Applications, 2023, 84 : 531 - 572
  • [2] Distributed stochastic gradient tracking methods with momentum acceleration for non-convex optimization
    Gao, Juan
    Liu, Xin-Wei
    Dai, Yu-Hong
    Huang, Yakui
    Gu, Junhua
    COMPUTATIONAL OPTIMIZATION AND APPLICATIONS, 2023, 84 (02) : 531 - 572
  • [3] Riemannian Stochastic Recursive Momentum Method for non-Convex Optimization
    Han, Andi
    Gao, Junbin
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2505 - 2511
  • [4] Adaptive Stochastic Gradient Descent Method for Convex and Non-Convex Optimization
    Chen, Ruijuan
    Tang, Xiaoquan
    Li, Xiuting
    FRACTAL AND FRACTIONAL, 2022, 6 (12)
  • [5] Stochastic momentum methods for non-convex learning without bounded assumptions
    Liang, Yuqing
    Liu, Jinlan
    Xu, Dongpo
    NEURAL NETWORKS, 2023, 165 : 830 - 845
  • [6] Decentralized non-convex online optimization with adaptive momentum estimation and quantized communication
    Lv, Yunshan
    Xiong, Hailing
    Zhang, Fuqing
    Dong, Shengying
    Dai, Xiangguang
    COMPLEX & INTELLIGENT SYSTEMS, 2025, 11 (04)
  • [7] Stochastic Successive Convex Approximation for Non-Convex Constrained Stochastic Optimization
    Liu, An
    Lau, Vincent K. N.
    Kananian, Borna
    IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2019, 67 (16) : 4189 - 4203
  • [8] Gradient Methods for Non-convex Optimization
    Jain, Prateek
    JOURNAL OF THE INDIAN INSTITUTE OF SCIENCE, 2019, 99 (02) : 247 - 256
  • [9] Gradient Methods for Non-convex Optimization
    Prateek Jain
    Journal of the Indian Institute of Science, 2019, 99 : 247 - 256
  • [10] Natasha: Faster Non-Convex Stochastic Optimization via Strongly Non-Convex Parameter
    Allen-Zhu, Zeyuan
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70