Convergence analysis of gradient descent stochastic algorithms

被引:53
|
作者
Shapiro, A [1 ]
Wardi, Y [1 ]
机构
[1] GEORGIA INST TECHNOL,SCH ELECT & COMP ENGN,ATLANTA,GA 30332
关键词
gradient descent; subdifferentials; uniform laws of large numbers; infinitesimal perturbation analysis; discrete event dynamic systems;
D O I
10.1007/BF02190104
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
This paper proves convergence of a sample-path based stochastic gradient-descent algorithm for optimizing expected-value performance measures in discrete event systems. The algorithm uses increasing precision at successive iterations, and it moves against the direction of a generalized gradient of the computed sample performance function. Two convergence results are established: one, for the case where the expected-value function is continuously differentiable; and the other, when that function is nondifferentiable but the sample performance functions are convex. The proofs are based on a version of the uniform law of large numbers which is provable for many discrete event systems where infinitesimal perturbation analysis is known to be strongly consistent.
引用
收藏
页码:439 / 454
页数:16
相关论文
共 50 条
  • [1] Convergence in High Probability of Distributed Stochastic Gradient Descent Algorithms
    Lu, Kaihong
    Wang, Hongxia
    Zhang, Huanshui
    Wang, Long
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2024, 69 (04) : 2189 - 2204
  • [2] Convergence analysis of distributed stochastic gradient descent with shuffling
    Meng, Qi
    Chen, Wei
    Wang, Yue
    Ma, Zhi-Ming
    Liu, Tie-Yan
    [J]. NEUROCOMPUTING, 2019, 337 : 46 - 57
  • [3] Convergence of Stochastic Gradient Descent for PCA
    Shamir, Ohad
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [4] A Tight Convergence Analysis for Stochastic Gradient Descent with Delayed Updates
    Arjevani, Yossi
    Shamir, Ohad
    Srebro, Nathan
    [J]. ALGORITHMIC LEARNING THEORY, VOL 117, 2020, 117 : 111 - 132
  • [5] Decentralized Asynchronous Stochastic Gradient Descent: Convergence Rate Analysis
    Bedi, Amrit Singh
    Pradhan, Hrusikesha
    Rajawat, Ketan
    [J]. 2018 INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND COMMUNICATIONS (SPCOM 2018), 2018, : 402 - 406
  • [6] Strong error analysis for stochastic gradient descent optimization algorithms
    Jentzen, Arnulf
    Kuckuck, Benno
    Neufeld, Ariel
    von Wurstemberger, Philippe
    [J]. IMA JOURNAL OF NUMERICAL ANALYSIS, 2021, 41 (01) : 455 - 492
  • [7] On the convergence and improvement of stochastic normalized gradient descent
    Shen-Yi ZHAO
    Yin-Peng XIE
    Wu-Jun LI
    [J]. Science China(Information Sciences), 2021, 64 (03) : 105 - 117
  • [8] Linear Convergence of Adaptive Stochastic Gradient Descent
    Xie, Yuege
    Wu, Xiaoxia
    Ward, Rachel
    [J]. arXiv, 2019,
  • [9] Convergence analysis of asynchronous stochastic recursive gradient algorithms
    Wang, Pengfei
    Zheng, Nenggan
    [J]. KNOWLEDGE-BASED SYSTEMS, 2022, 252
  • [10] On the Convergence of Stochastic Gradient Descent with Adaptive Stepsizes
    Li, Xiaoyu
    Orabona, Francesco
    [J]. 22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89