On the Saturation Phenomenon of Stochastic Gradient Descent for Linear Inverse Problems*

被引:4
|
作者
Jin, Bangti [1 ]
Zhou, Zehui [2 ]
Zou, Jun [2 ]
机构
[1] UCL, Dept Comp Sci, London WC1E 6BT, England
[2] Chinese Univ Hong Kong, Dept Math, Shatin, Hong Kong, Peoples R China
来源
基金
英国工程与自然科学研究理事会;
关键词
  stochastic gradient descent; regularizing property; convergence rate; saturation; inverse problems; APPROXIMATION; CONVERGENCE;
D O I
10.1137/20M1374456
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Stochastic gradient descent (SGD) is a promising method for solving large-scale inverse problems due to its excellent scalability with respect to data size. The current mathematical theory in the lens of regularization theory predicts that SGD with a polynomially decaying stepsize schedule may suffer from an undesirable saturation phenomenon; i.e., the convergence rate does not further improve with the solution regularity index when it is beyond a certain range. In this work, we present a refined convergence rate analysis of SGD and prove that saturation actually does not occur if the initial stepsize of the schedule is sufficiently small. Several numerical experiments are provided to complement the analysis.
引用
收藏
页码:1553 / 1588
页数:36
相关论文
共 50 条
  • [31] Forecasting with imperfect models, dynamically constrained inverse problems, and gradient descent algorithms
    Judd, Kevin
    PHYSICA D-NONLINEAR PHENOMENA, 2008, 237 (02) : 216 - 232
  • [32] Quantifying Model Uncertainty in Inverse Problems via Bayesian Deep Gradient Descent
    Barbano, Riccardo
    Zhang, Chen
    Arridge, Simon
    Jin, Bangti
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 1392 - 1399
  • [33] Lévy Gradient Descent: Augmented Random Search for Geophysical Inverse Problems
    Yi Zhang
    Yixian Xu
    Bo Yang
    Surveys in Geophysics, 2021, 42 : 899 - 921
  • [34] Sparse Stochastic Processes and Discretization of Linear Inverse Problems
    Bostan, Emrah
    Kamilov, Ulugbek S.
    Nilchian, Masih
    Unser, Michael
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2013, 22 (07) : 2699 - 2710
  • [35] Recovery Guarantees of Unsupervised Neural Networks for Inverse Problems trained with Gradient Descent
    Buskulic, Nathan
    Queau, Yvain
    Fadili, Jalal
    32ND EUROPEAN SIGNAL PROCESSING CONFERENCE, EUSIPCO 2024, 2024, : 1806 - 1810
  • [36] THE LIMITATION AND PRACTICAL ACCELERATION OF STOCHASTIC GRADIENT ALGORITHMS IN INVERSE PROBLEMS
    Tang, Junqi
    Egiazarian, Karen
    Davies, Mike
    2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 7680 - 7684
  • [37] Stochastic Recursive Gradient Descent Ascent for Stochastic Nonconvex-Strongly-Concave Minimax Problems
    Luo, Luo
    Ye, Haishan
    Huang, Zhichao
    Zhang, Tong
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [38] Machine learning for inverse lithography: using stochastic gradient descent for robust photomask synthesis
    Jia, Ningning
    Lam, Edmund Y.
    JOURNAL OF OPTICS, 2010, 12 (04)
  • [39] Descent methods for inverse problems
    Knowles, I
    NONLINEAR ANALYSIS-THEORY METHODS & APPLICATIONS, 2001, 47 (05) : 3235 - 3245
  • [40] Byzantine Fault-Tolerant Parallelized Stochastic Gradient Descent for Linear Regression
    Gupta, Nirupam
    Vaidya, Nitin H.
    2019 57TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2019, : 415 - 420