Almost sure convergence rates of stochastic proximal gradient descent algorithm

被引:0
|
作者
Liang, Yuqing [1 ]
Xu, Dongpo [1 ]
机构
[1] Northeast Normal Univ, Sch Math & Stat, Key Lab Appl Stat MOE, Changchun, Peoples R China
基金
中国国家自然科学基金;
关键词
Stochastic proximal gradient descent algorithm; almost sure convergence; last-iterate convergence; convexity; strong convexity; OPTIMIZATION;
D O I
10.1080/02331934.2023.2230976
中图分类号
C93 [管理学]; O22 [运筹学];
学科分类号
070105 ; 12 ; 1201 ; 1202 ; 120202 ;
摘要
Stochastic proximal gradient descent (Prox-SGD) is a standard optimization algorithm for solving stochastic composite opti-mization problems in machine learning. However, the existing convergence rate analysis of Prox-SGD is mostly focused on convergence in expectation. To this end, we provide an almost sure convergence rate analysis of Prox-SGD in two optimiza-tion settings: convexity and strong convexity, showing that any implementation of the stochastic algorithm will converge with probability one. Specifically, we show that the average-iterates almost sure convergence rate of the function values v for Prox-SGD is arbitrarily close to the optimal rate o(1/v T) in the convex setting. For strongly convex objective functions, we obtain the almost sure convergence rate of the iterative sequence, which is also arbitrarily close to the optimal rate o(1/T). Last but not least, we establish the more challeng-ing last-iterate convergence rate of the function values on convex problems, which contrasts with existing results on convergence for a weighted average.
引用
收藏
页码:2413 / 2446
页数:34
相关论文
共 50 条
  • [1] Further results on almost sure convergence of the stochastic gradient algorithm
    Radenkovic, MS
    [J]. INTERNATIONAL JOURNAL OF ADAPTIVE CONTROL AND SIGNAL PROCESSING, 1999, 13 (03) : 161 - 181
  • [2] On the Almost Sure Convergence of Stochastic Gradient Descent in Non-Convex Problems
    Mertikopoulos, Panayotis
    Hallak, Nadav
    Kavis, Ali
    Cevher, Volkan
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [3] Almost sure convergence of randomised-difference descent algorithm for stochastic convex optimisation
    Geng, Xiaoxue
    Huang, Gao
    Zhao, Wenxiao
    [J]. IET CONTROL THEORY AND APPLICATIONS, 2021, 15 (17): : 2183 - 2194
  • [4] Convergence of Stochastic Proximal Gradient Algorithm
    Lorenzo Rosasco
    Silvia Villa
    Bằng Công Vũ
    [J]. Applied Mathematics & Optimization, 2020, 82 : 891 - 917
  • [5] Convergence of Stochastic Proximal Gradient Algorithm
    Rosasco, Lorenzo
    Villa, Silvia
    Vu, Bing Cong
    [J]. APPLIED MATHEMATICS AND OPTIMIZATION, 2020, 82 (03): : 891 - 917
  • [6] Estimating the geometric median in Hilbert spaces with stochastic gradient algorithms: Lp and almost sure rates of convergence
    Godichon-Baggioni, Antoine
    [J]. JOURNAL OF MULTIVARIATE ANALYSIS, 2016, 146 : 209 - 222
  • [7] Almost sure convergence of stochastic gradient processes with matrix step sizes
    Monnez, JM
    [J]. STATISTICS & PROBABILITY LETTERS, 2006, 76 (05) : 531 - 536
  • [8] Lp and almost sure rates of convergence of averaged stochastic gradient algorithms: locally strongly convex objective
    Godichon-Baggioni, Antoine
    [J]. ESAIM-PROBABILITY AND STATISTICS, 2019, 23 : 841 - 873
  • [9] An Almost Sure Convergence Analysis of Zeroth-Order Mirror Descent Algorithm
    Paul, Anik Kumar
    Mahindrakar, Arun D.
    Kalaimani, Rachel K.
    [J]. 2023 AMERICAN CONTROL CONFERENCE, ACC, 2023, : 855 - 860
  • [10] Fast Convergence Stochastic Parallel Gradient Descent Algorithm
    Hu Dongting
    Shen Wen
    Ma Wenchao
    Liu Xinyu
    Su Zhouping
    Zhu Huaxin
    Zhang Xiumei
    Que Lizhi
    Zhu Zhuowei
    Zhang Yixin
    Chen Guoqing
    Hu Lifa
    [J]. LASER & OPTOELECTRONICS PROGRESS, 2019, 56 (12)