On the Convergence of Stochastic Gradient Descent for Linear Inverse Problems in Banach Spaces

被引:2
|
作者
Jin, Bangti [1 ]
Kereta, Zeljko [2 ]
机构
[1] Chinese Univ Hong Kong, Dept Math, Shatin, Hong Kong, Peoples R China
[2] UCL, Dept Comp Sci, Gower St, London WC1E 6BT, England
来源
SIAM JOURNAL ON IMAGING SCIENCES | 2023年 / 16卷 / 02期
基金
英国工程与自然科学研究理事会;
关键词
stochastic gradient descent; Banach spaces; linear inverse problems; convergence rate; regularizing property; almost sure convergence; ILL-POSED PROBLEMS; REGULARIZATION; OPTIMIZATION; PARAMETERS; KACZMARZ; CHOICE;
D O I
10.1137/22M1518542
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this work we consider stochastic gradient descent (SGD) for solving linear inverse problems in Banach spaces. SGD and its variants have been established as one of the most successful optimization methods in machine learning, imaging, and signal processing, to name a few. At each iteration SGD uses a single datum, or a small subset of data, resulting in highly scalable methods that are very attractive for large-scale inverse problems. Nonetheless, the theoretical analysis of SGD-based approaches for inverse problems has thus far been largely limited to Euclidean and Hilbert spaces. In this work we present a novel convergence analysis of SGD for linear inverse problems in general Banach spaces: we show the almost sure convergence of the iterates to the minimum norm solution and establish the regularizing property for suitable a priori stopping criteria. Numerical results are also presented to illustrate features of the approach.
引用
收藏
页码:671 / 705
页数:35
相关论文
共 50 条
  • [21] Tight Nonparametric Convergence Rates for Stochastic Gradient Descent under the Noiseless Linear Model
    Berthier, Raphael
    Bach, Francis
    Gaillard, Pierre
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [22] Global Convergence of Stochastic Gradient Descent for Some Non-convex Matrix Problems
    De Sa, Christopher
    Olukotun, Kunle
    Re, Christopher
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 37, 2015, 37 : 2332 - 2341
  • [23] Generalized Convergence Rates Results for Linear Inverse Problems in Hilbert Spaces
    Andreev, Roman
    Elbau, Peter
    de Hoop, Maarten V.
    Qiu, Lingyun
    Scherzer, Otmar
    NUMERICAL FUNCTIONAL ANALYSIS AND OPTIMIZATION, 2015, 36 (05) : 549 - 566
  • [24] Optimal Convergence Rates Results for Linear Inverse Problems in Hilbert Spaces
    Albani, V.
    Elbau, P.
    de Hoop, M. V.
    Scherzer, O.
    NUMERICAL FUNCTIONAL ANALYSIS AND OPTIMIZATION, 2016, 37 (05) : 521 - 540
  • [25] A generic convergence theorem for continuous descent methods in Banach spaces
    Aizicovici, Sergiu
    Reich, Simeon
    Zaslavski, Alexander J.
    DYNAMICS OF CONTINUOUS DISCRETE AND IMPULSIVE SYSTEMS-SERIES A-MATHEMATICAL ANALYSIS, 2007, 14 (01): : 137 - 145
  • [26] Accelerated Landweber iteration with convex penalty for linear inverse problems in Banach spaces
    Hegland, Markus
    Jin, Qinian
    Wang, Wei
    APPLICABLE ANALYSIS, 2015, 94 (03) : 524 - 547
  • [27] Convergence of Stochastic Gradient Descent in Deep Neural Network
    Bai-cun Zhou
    Cong-ying Han
    Tian-de Guo
    Acta Mathematicae Applicatae Sinica, English Series, 2021, 37 : 126 - 136
  • [28] Optimized convergence of stochastic gradient descent by weighted averaging
    Hagedorn, Melinda
    Jarre, Florian
    OPTIMIZATION METHODS & SOFTWARE, 2024, 39 (04): : 699 - 724
  • [29] Convergence analysis of distributed stochastic gradient descent with shuffling
    Meng, Qi
    Chen, Wei
    Wang, Yue
    Ma, Zhi-Ming
    Liu, Tie-Yan
    NEUROCOMPUTING, 2019, 337 : 46 - 57
  • [30] Fast Convergence Stochastic Parallel Gradient Descent Algorithm
    Hu Dongting
    Shen Wen
    Ma Wenchao
    Liu Xinyu
    Su Zhouping
    Zhu Huaxin
    Zhang Xiumei
    Que Lizhi
    Zhu Zhuowei
    Zhang Yixin
    Chen Guoqing
    Hu Lifa
    LASER & OPTOELECTRONICS PROGRESS, 2019, 56 (12)