On the Saturation Phenomenon of Stochastic Gradient Descent for Linear Inverse Problems*

被引:4
|
作者
Jin, Bangti [1 ]
Zhou, Zehui [2 ]
Zou, Jun [2 ]
机构
[1] UCL, Dept Comp Sci, London WC1E 6BT, England
[2] Chinese Univ Hong Kong, Dept Math, Shatin, Hong Kong, Peoples R China
来源
基金
英国工程与自然科学研究理事会;
关键词
  stochastic gradient descent; regularizing property; convergence rate; saturation; inverse problems; APPROXIMATION; CONVERGENCE;
D O I
10.1137/20M1374456
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Stochastic gradient descent (SGD) is a promising method for solving large-scale inverse problems due to its excellent scalability with respect to data size. The current mathematical theory in the lens of regularization theory predicts that SGD with a polynomially decaying stepsize schedule may suffer from an undesirable saturation phenomenon; i.e., the convergence rate does not further improve with the solution regularity index when it is beyond a certain range. In this work, we present a refined convergence rate analysis of SGD and prove that saturation actually does not occur if the initial stepsize of the schedule is sufficiently small. Several numerical experiments are provided to complement the analysis.
引用
收藏
页码:1553 / 1588
页数:36
相关论文
共 50 条
  • [21] Stochastic gradient descent for linear systems with sequential matrix entry accumulation
    Mukhopadhyay, Samrat
    1600, Elsevier B.V., Netherlands (171):
  • [22] Unforgeability in Stochastic Gradient Descent
    Baluta, Teodora
    Nikolic, Ivica
    Jain, Racchit
    Aggarwal, Divesh
    Saxena, Prateek
    PROCEEDINGS OF THE 2023 ACM SIGSAC CONFERENCE ON COMPUTER AND COMMUNICATIONS SECURITY, CCS 2023, 2023, : 1138 - 1152
  • [23] Preconditioned Stochastic Gradient Descent
    Li, Xi-Lin
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2018, 29 (05) : 1454 - 1466
  • [24] Stochastic Reweighted Gradient Descent
    El Hanchi, Ayoub
    Stephens, David A.
    Maddison, Chris J.
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [25] ON THE CONVERGENCE OF STOCHASTIC GRADIENT DESCENT FOR NONLINEAR ILL-POSED PROBLEMS
    Jin, Bangti
    Zhou, Zehui
    Zou, Jun
    SIAM JOURNAL ON OPTIMIZATION, 2020, 30 (02) : 1421 - 1450
  • [26] Stochastic gradient descent tricks
    Bottou, Léon
    Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2012, 7700 LECTURE NO : 421 - 436
  • [27] Byzantine Stochastic Gradient Descent
    Alistarh, Dan
    Allen-Zhu, Zeyuan
    Li, Jerry
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [28] The Gradient Descent Method for the Convexification to Solve Boundary Value Problems of Quasi-Linear PDEs and a Coefficient Inverse Problem
    Le, Thuy T.
    Nguyen, Loc H.
    JOURNAL OF SCIENTIFIC COMPUTING, 2022, 91 (03)
  • [29] The Gradient Descent Method for the Convexification to Solve Boundary Value Problems of Quasi-Linear PDEs and a Coefficient Inverse Problem
    Thuy T. Le
    Loc H. Nguyen
    Journal of Scientific Computing, 2022, 91
  • [30] A descent method for computing the Tikhonov regularized solution of linear inverse problems
    Zama, F
    Piccolomini, EL
    Landi, G
    IMAGE RECONSTRUCTION FROM INCOMPLETE DATA III, 2004, 5562 : 152 - 160