A momentum accelerated stochastic method and its application on policy search problems

被引:0
|
作者
Boou Jiang [1 ]
Ya-xiang Yuan [2 ]
机构
[1] LSEC,
[2] ICMSEC,undefined
[3] AMSS,undefined
[4] Chniese Academy of Sciences,undefined
[5] University of Chinese Academy of Sciences,undefined
关键词
Stochastic algorithm; Non-convex optimization; Reinforcement learning; 65K05; 90C15; 90C25; 90C30;
D O I
10.1007/s00521-024-10883-y
中图分类号
学科分类号
摘要
With the dramatic increase in model complexity and problem scales in the machine learning area, researches on the first-order stochastic methods and its accelerated variants for non-convex problems have attracted wide research interest. However, most works on convergence analysis of accelerated methods focus on general convex or strongly convex objective functions. In this paper, we consider an accelerated scheme coming from dynamic systems and ordinary differential equations, which has a simpler and more direct form than the traditional scheme. We construct auxiliary sequences of iteration points as analysis tools, which can be interpreted as extension of Nesterov’s estimate sequence in non-convex case. We analyze the convergence property under different cases when momentum parameters are fixed or varying over iterations. For non-smooth and general convex objective functions, we give a relaxed step-size requirement to ensure convergence. For the non-convex policy search problem in classical reinforcement learning, we propose an accelerated stochastic policy gradient method with restart technique and construct numerical experiments to verify its effectiveness.
引用
收藏
页码:5957 / 5973
页数:16
相关论文
共 50 条
  • [1] An accelerated distributed stochastic gradient method with momentum
    Huang, Kun
    Pu, Shi
    Nedic, Angelia
    MATHEMATICAL PROGRAMMING, 2025,
  • [2] A method of accelerated statistical simulation and its application in the problems with inherent error
    Nekrasov S.A.
    Journal of Applied and Industrial Mathematics, 2017, 11 (2) : 244 - 251
  • [3] A Stochastic Momentum Accelerated Quasi-Newton Method for Neural Networks
    Indrapriyadarsini, S.
    Mahboubi, Shahrzad
    Ninomiya, Hiroshi
    Kamio, Takeshi
    Asai, Hideki
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 12973 - 12974
  • [4] A Nonmonotone Line Search Method for Stochastic Optimization Problems
    Krejic, Natasa
    Loncar, Sanja
    FILOMAT, 2018, 32 (19) : 6799 - 6807
  • [5] SMOOTHING PROJECTED GRADIENT METHOD AND ITS APPLICATION TO STOCHASTIC LINEAR COMPLEMENTARITY PROBLEMS
    Zhang, Chao
    Chen, Xiaojun
    SIAM JOURNAL ON OPTIMIZATION, 2009, 20 (02) : 627 - 649
  • [6] A dynamic stochastic search algorithm for high-dimensional optimization problems and its application to feature selection
    Liu, Qi
    Liu, Mengxue
    Wang, Fengde
    Xiao, Wensheng
    KNOWLEDGE-BASED SYSTEMS, 2022, 244
  • [7] An accelerated stochastic variance-reduced method for machine learning problems
    Yang, Zhuang
    Chen, Zengping
    Wang, Cheng
    KNOWLEDGE-BASED SYSTEMS, 2020, 198
  • [8] EFFECTIVE METHOD FOR ACCELERATED SEARCH FOR PROBLEMS OF TOPOLOGICAL OPTIMIZATION OF INFORMATION NETWORKS
    ZAICHENKO, YP
    ANDROPOV, VM
    AVTOMATIKA, 1986, (01): : 16 - 23
  • [9] Scheduled Restart Momentum for Accelerated Stochastic Gradient Descent
    Wang, Bao
    Nguyen, Tan
    Sun, Tao
    Bertozzi, Andrea L.
    Baraniuk, Richard G.
    Osher, Stanley J.
    SIAM JOURNAL ON IMAGING SCIENCES, 2022, 15 (02): : 738 - 761
  • [10] Stochastic Fractal Search Algorithm and its Application in Path Planning
    Li, Wenguang
    Un, Shiyu
    Li, Jianzeng
    Hu, Yongjiang
    2018 IEEE CSAA GUIDANCE, NAVIGATION AND CONTROL CONFERENCE (CGNCC), 2018,