On the Saturation Phenomenon of Stochastic Gradient Descent for Linear Inverse Problems*

被引:4
|
作者
Jin, Bangti [1 ]
Zhou, Zehui [2 ]
Zou, Jun [2 ]
机构
[1] UCL, Dept Comp Sci, London WC1E 6BT, England
[2] Chinese Univ Hong Kong, Dept Math, Shatin, Hong Kong, Peoples R China
来源
基金
英国工程与自然科学研究理事会;
关键词
  stochastic gradient descent; regularizing property; convergence rate; saturation; inverse problems; APPROXIMATION; CONVERGENCE;
D O I
10.1137/20M1374456
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Stochastic gradient descent (SGD) is a promising method for solving large-scale inverse problems due to its excellent scalability with respect to data size. The current mathematical theory in the lens of regularization theory predicts that SGD with a polynomially decaying stepsize schedule may suffer from an undesirable saturation phenomenon; i.e., the convergence rate does not further improve with the solution regularity index when it is beyond a certain range. In this work, we present a refined convergence rate analysis of SGD and prove that saturation actually does not occur if the initial stepsize of the schedule is sufficiently small. Several numerical experiments are provided to complement the analysis.
引用
收藏
页码:1553 / 1588
页数:36
相关论文
共 50 条
  • [1] STOCHASTIC GRADIENT DESCENT FOR LINEAR INVERSE PROBLEMS IN HILBERT SPACES
    Lu, Shuai
    Mathe, Peter
    MATHEMATICS OF COMPUTATION, 2022, 91 (336) : 1763 - 1788
  • [2] On the Convergence of Stochastic Gradient Descent for Linear Inverse Problems in Banach Spaces
    Jin, Bangti
    Kereta, Zeljko
    SIAM JOURNAL ON IMAGING SCIENCES, 2023, 16 (02): : 671 - 705
  • [3] An analysis of stochastic variance reduced gradient for linear inverse problems *
    Jin, Bangti
    Zhou, Zehui
    Zou, Jun
    INVERSE PROBLEMS, 2022, 38 (02)
  • [4] Linear Convergence of Adaptive Stochastic Gradient Descent
    Xie, Yuege
    Wu, Xiaoxia
    Ward, Rachel
    arXiv, 2019,
  • [5] Linear Convergence of Adaptive Stochastic Gradient Descent
    Xie, Yuege
    Wu, Xiaoxia
    Ward, Rachel
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 108, 2020, 108
  • [6] Stochastic Gradient Descent for Linear Systems with Missing Data
    Ma, Anna
    Needell, Deanna
    NUMERICAL MATHEMATICS-THEORY METHODS AND APPLICATIONS, 2019, 12 (01) : 1 - 20
  • [7] THE ALTERNATING DESCENT CONDITIONAL GRADIENT METHOD FOR SPARSE INVERSE PROBLEMS
    Boyd, Nicholas
    Schiebinger, Geoffrey
    Recht, Benjamin
    SIAM JOURNAL ON OPTIMIZATION, 2017, 27 (02) : 616 - 639
  • [8] The Alternating Descent Conditional Gradient Method for Sparse Inverse Problems
    Boyd, Nicholas
    Schiebinger, Geoffrey
    Recht, Benjamin
    2015 IEEE 6TH INTERNATIONAL WORKSHOP ON COMPUTATIONAL ADVANCES IN MULTI-SENSOR ADAPTIVE PROCESSING (CAMSAP), 2015, : 57 - 60
  • [9] STOCHASTIC GRADIENT DESCENT FOR ROBUST INVERSE PHOTOMASK SYNTHESIS IN OPTICAL LITHOGRAPHY
    Jia, Ningning
    Lam, Edmund Y.
    2010 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, 2010, : 4173 - 4176
  • [10] Stochastic Gradient Descent Variants for Corrupted Systems of Linear Equations
    Haddock, Jamie
    Needell, Deanna
    Rebrova, Elizaveta
    Swartworth, William
    2020 54TH ANNUAL CONFERENCE ON INFORMATION SCIENCES AND SYSTEMS (CISS), 2020, : 348 - 353