Convergence of Stochastic Gradient Descent for PCA

被引:0
|
作者
Shamir, Ohad [1 ]
机构
[1] Weizmann Inst Sci, Rehovot, Israel
基金
以色列科学基金会;
关键词
ALGORITHMS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider the problem of principal component analysis (PCA) in a streaming stochastic setting, where our goal is to find a direction of approximate maximal variance, based on a stream of i.i.d. data points in I': d. A simple and computationally cheap algorithm for this is stochastic gradient descent (SGD), which incrementally updates its estimate based on each new data point. However, due to the non-convex nature of the problem, analyzing its performance has been a challenge. In particular, existing guarantees rely on a non-trivial eigengap assumption on the covariance matrix, which is intuitively unnecessary. In this paper, we provide (to the best of our knowledge) the first eigengap-free convergence guarantees for SGD in the context of PCA. This also partially resolves an open problem posed in (Hardt & Price, 2014). Moreover, under an eigengap assumption, we show that the same techniques lead to new SGD convergence guarantees with better dependence on the eigengap.
引用
收藏
页数:9
相关论文
共 50 条
  • [21] Image Alignment by Online Robust PCA via Stochastic Gradient Descent
    Song, Wenjie
    Zhu, Jianke
    Li, Yang
    Chen, Chun
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2016, 26 (07) : 1241 - 1250
  • [22] Convergence diagnostics for stochastic gradient descent with constant learning rate
    Chee, Jerry
    Toulis, Panos
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 84, 2018, 84
  • [23] A simplified convergence theory for Byzantine resilient stochastic gradient descent
    Roberts, Lindon
    Smyth, Edward
    [J]. EURO JOURNAL ON COMPUTATIONAL OPTIMIZATION, 2022, 10
  • [24] A Tight Convergence Analysis for Stochastic Gradient Descent with Delayed Updates
    Arjevani, Yossi
    Shamir, Ohad
    Srebro, Nathan
    [J]. ALGORITHMIC LEARNING THEORY, VOL 117, 2020, 117 : 111 - 132
  • [25] Convergence in High Probability of Distributed Stochastic Gradient Descent Algorithms
    Lu, Kaihong
    Wang, Hongxia
    Zhang, Huanshui
    Wang, Long
    [J]. IEEE TRANSACTIONS ON AUTOMATIC CONTROL, 2024, 69 (04) : 2189 - 2204
  • [26] Decentralized Asynchronous Stochastic Gradient Descent: Convergence Rate Analysis
    Bedi, Amrit Singh
    Pradhan, Hrusikesha
    Rajawat, Ketan
    [J]. 2018 INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND COMMUNICATIONS (SPCOM 2018), 2018, : 402 - 406
  • [27] Fast Convergence for Stochastic and Distributed Gradient Descent in the Interpolation Limit
    Mitra, Partha P.
    [J]. 2018 26TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2018, : 1890 - 1894
  • [28] Almost sure convergence rates of stochastic proximal gradient descent algorithm
    Liang, Yuqing
    Xu, Dongpo
    [J]. OPTIMIZATION, 2024, 73 (08) : 2413 - 2446
  • [29] Convergence Analysis of Accelerated Stochastic Gradient Descent Under the Growth Condition
    Chen, You-Lin
    Na, Sen
    Kolar, Mladen
    [J]. MATHEMATICS OF OPERATIONS RESEARCH, 2023,
  • [30] Distributed Stochastic Gradient Descent: Nonconvexity, Nonsmoothness, and Convergence to Local Minima
    Swenson, Brian
    Murray, Ryan
    Poor, H. Vincent
    Kar, Soummya
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2022, 23