Thinking Outside the Ball: Optimal Learning with Gradient Descent for Generalized Linear Stochastic Convex Optimization

被引:0
|
作者
Amir, Idan [1 ]
Livni, Roi [1 ]
Srebro, Nathan [2 ]
机构
[1] Tel Aviv Univ, Tel Aviv, Israel
[2] Toyota Technol Inst, Chicago, IL USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider linear prediction with a convex Lipschitz loss, or more generally, stochastic convex optimization problems of generalized linear form, i.e. where each instantaneous loss is a scalar convex function of a linear function. We show that in this setting, early stopped Gradient Descent (GD), without any explicit regularization or projection, ensures excess error at most epsilon (compared to the best possible with unit Euclidean norm) with an optimal, up to logarithmic factors, sample complexity of (O) over tilde (1/epsilon(2)) and only (O) over tilde (1/epsilon(2)) iterations. This contrasts with general stochastic convex optimization, where (O) over tilde (1/epsilon(4)) iterations are needed Amir et al. [2]. The lower iteration complexity is ensured by leveraging uniform convergence rather than stability. But instead of uniform convergence in a norm ball, which we show can guarantee suboptimal learning using Theta(1/epsilon(4)) samples, we rely on uniform convergence in a distribution-dependent ball.
引用
收藏
页数:12
相关论文
共 50 条
  • [41] On the Convergence of (Stochastic) Gradient Descent with Extrapolation for Non-Convex Minimization
    Xu, Yi
    Yuan, Zhuoning
    Yang, Sen
    Jin, Rong
    Yang, Tianbao
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 4003 - 4009
  • [42] Nonconvex Stochastic Scaled Gradient Descent and Generalized Eigenvector Problems
    Li, Chris Junchi
    Jordan, Michael I.
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2023, 216 : 1230 - 1240
  • [43] BAYESIAN STOCHASTIC GRADIENT DESCENT FOR STOCHASTIC OPTIMIZATION WITH STREAMING INPUT DATA
    Liu, Tianyi
    Lin, Yifan
    Zhou, Enlu
    SIAM JOURNAL ON OPTIMIZATION, 2024, 34 (01) : 389 - 418
  • [44] Algorithms of Inertial Mirror Descent in Convex Problems of Stochastic Optimization
    Nazin, A. V.
    AUTOMATION AND REMOTE CONTROL, 2018, 79 (01) : 78 - 88
  • [45] Algorithms of Inertial Mirror Descent in Convex Problems of Stochastic Optimization
    A. V. Nazin
    Automation and Remote Control, 2018, 79 : 78 - 88
  • [46] Stochastic Mirror Descent for Convex Optimization with Consensus Constraints\ast
    Borovykh, A.
    Kantas, N.
    Parpas, P.
    Pavliotis, G. A.
    SIAM JOURNAL ON APPLIED DYNAMICAL SYSTEMS, 2024, 23 (03): : 2208 - 2241
  • [47] Fast Stochastic Kalman Gradient Descent for Reinforcement Learning
    Totaro, Simone
    Jonsson, Anders
    LEARNING FOR DYNAMICS AND CONTROL, VOL 144, 2021, 144
  • [48] Stochastic Gradient Descent and Its Variants in Machine Learning
    Netrapalli, Praneeth
    JOURNAL OF THE INDIAN INSTITUTE OF SCIENCE, 2019, 99 (02) : 201 - 213
  • [49] Towards Learning Stochastic Population Models by Gradient Descent
    Kreikemeyer, Justin N.
    Andelfinger, Philipp
    Uhrmacher, Adelinde M.
    PROCEEDINGS OF THE 38TH ACM SIGSIM INTERNATIONAL CONFERENCE ON PRINCIPLES OF ADVANCED DISCRETE SIMULATION, ACM SIGSIM-PADS 2024, 2024, : 88 - 92
  • [50] Stochastic Gradient Descent with Polyak's Learning Rate
    Prazeres, Mariana
    Oberman, Adam M.
    JOURNAL OF SCIENTIFIC COMPUTING, 2021, 89 (01)