Convergence of Stochastic Gradient Descent for PCA

被引:0
|
作者
Shamir, Ohad [1 ]
机构
[1] Weizmann Inst Sci, Rehovot, Israel
基金
以色列科学基金会;
关键词
ALGORITHMS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider the problem of principal component analysis (PCA) in a streaming stochastic setting, where our goal is to find a direction of approximate maximal variance, based on a stream of i.i.d. data points in I': d. A simple and computationally cheap algorithm for this is stochastic gradient descent (SGD), which incrementally updates its estimate based on each new data point. However, due to the non-convex nature of the problem, analyzing its performance has been a challenge. In particular, existing guarantees rely on a non-trivial eigengap assumption on the covariance matrix, which is intuitively unnecessary. In this paper, we provide (to the best of our knowledge) the first eigengap-free convergence guarantees for SGD in the context of PCA. This also partially resolves an open problem posed in (Hardt & Price, 2014). Moreover, under an eigengap assumption, we show that the same techniques lead to new SGD convergence guarantees with better dependence on the eigengap.
引用
收藏
页数:9
相关论文
共 50 条
  • [31] A SHARP CONVERGENCE RATE FOR A MODEL EQUATION OF THE ASYNCHRONOUS STOCHASTIC GRADIENT DESCENT
    Zhu, Yuhua
    Ying, Lexing
    [J]. COMMUNICATIONS IN MATHEMATICAL SCIENCES, 2021, 19 (03) : 851 - 863
  • [32] SUBLINEAR CONVERGENCE OF A TAMED STOCHASTIC GRADIENT DESCENT METHOD IN HILBERT SPACE
    Eisenmann, Monika
    Stillfjord, Tony
    [J]. SIAM JOURNAL ON OPTIMIZATION, 2022, 32 (03) : 1642 - 1667
  • [33] Convergence Rates of Stochastic Gradient Descent under Infinite Noise Variance
    Wang, Hongjian
    Gurbuzbalaban, Mert
    Zhu, Lingjiong
    Simsekli, Umut
    Erdogdu, Murat A.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [34] Convergence of Hyperbolic Neural Networks Under Riemannian Stochastic Gradient Descent
    Whiting, Wes
    Wang, Bao
    Xin, Jack
    [J]. COMMUNICATIONS ON APPLIED MATHEMATICS AND COMPUTATION, 2024, 6 (02) : 1175 - 1188
  • [35] Local Stochastic Gradient Descent Ascent: Convergence Analysis and Communication Efficiency
    Deng, Yuyang
    Mandavi, Mehrdad
    [J]. 24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130
  • [36] On the Convergence of Stochastic Gradient Descent for Linear Inverse Problems in Banach Spaces
    Jin, Bangti
    Kereta, Zeljko
    [J]. SIAM JOURNAL ON IMAGING SCIENCES, 2023, 16 (02): : 671 - 705
  • [37] On the Convergence of (Stochastic) Gradient Descent with Extrapolation for Non-Convex Minimization
    Xu, Yi
    Yuan, Zhuoning
    Yang, Sen
    Jin, Rong
    Yang, Tianbao
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 4003 - 4009
  • [38] Stochastic Gradient Descent with Exponential Convergence Rates of Expected Classification Errors
    Nitanda, Atsushi
    Suzuki, Taiji
    [J]. 22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [39] SUBLINEAR CONVERGENCE OF A TAMED STOCHASTIC GRADIENT DESCENT METHOD IN HILBERT SPACE
    Eisenmann, Monika
    Stillfjord, Tony
    [J]. SIAM Journal on Computing, 2022, 51 (04)
  • [40] On the Convergence of Stochastic Gradient Descent with Bandwidth-based Step Size
    Wang, Xiaoyu
    Yuan, Ya-xiang
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24