Large-Scale Machine Learning with Stochastic Gradient Descent

被引:3742
|
作者
Bottou, Leon [1 ]
机构
[1] NEC Labs Amer, Princeton, NJ 08542 USA
来源
COMPSTAT'2010: 19TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL STATISTICS | 2010年
关键词
stochastic gradient descent; online learning; efficiency;
D O I
10.1007/978-3-7908-2604-3_16
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
During the last decade, the data sizes have grown faster than the speed of processors. In this context, the capabilities of statistical machine learning methods is limited by the computing time rather than the sample size. A more precise analysis uncovers qualitatively different tradeoffs for the case of small-scale and large-scale learning problems. The large-scale case involves the computational complexity of the underlying optimization algorithm in non-trivial ways. Unlikely optimization algorithms such as stochastic gradient descent show amazing performance for large-scale problems. In particular, second order stochastic gradient and averaged stochastic gradient are asymptotically efficient after a single pass on the training set.
引用
收藏
页码:177 / 186
页数:10
相关论文
共 50 条
  • [21] Improved Powered Stochastic Optimization Algorithms for Large-Scale Machine Learning
    Yang, Zhuang
    JOURNAL OF MACHINE LEARNING RESEARCH, 2023, 24
  • [22] Accelerated Variance Reduction Stochastic ADMM for Large-Scale Machine Learning
    Liu, Yuanyuan
    Shang, Fanhua
    Liu, Hongying
    Kong, Lin
    Jiao, Licheng
    Lin, Zhouchen
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (12) : 4242 - 4255
  • [23] The Powerball Method With Biased Stochastic Gradient Estimation for Large-Scale Learning Systems
    Yang, Zhuang
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2024,
  • [24] Value function gradient learning for large-scale multistage stochastic programming problems
    Lee, Jinkyu
    Bae, Sanghyeon
    Kim, Woo Chang
    Lee, Yongjae
    EUROPEAN JOURNAL OF OPERATIONAL RESEARCH, 2023, 308 (01) : 321 - 335
  • [25] RECENT TRENDS IN STOCHASTIC GRADIENT DESCENT FOR MACHINE LEARNING AND BIG DATA
    Newton, David
    Pasupathy, Raghu
    Yousefian, Farzad
    2018 WINTER SIMULATION CONFERENCE (WSC), 2018, : 366 - 380
  • [26] An online conjugate gradient algorithm for large-scale data analysis in machine learning
    Xue, Wei
    Wan, Pengcheng
    Li, Qiao
    Zhong, Ping
    Yu, Gaohang
    Tao, Tao
    AIMS MATHEMATICS, 2021, 6 (02): : 1515 - 1537
  • [27] Stochastic trust region inexact Newton method for large-scale machine learning
    Vinod Kumar Chauhan
    Anuj Sharma
    Kalpana Dahiya
    International Journal of Machine Learning and Cybernetics, 2020, 11 : 1541 - 1555
  • [28] A Survey on Large-Scale Machine Learning
    Wang, Meng
    Fu, Weijie
    He, Xiangnan
    Hao, Shijie
    Wu, Xindong
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2022, 34 (06) : 2574 - 2594
  • [29] An Overview of Stochastic Quasi-Newton Methods for Large-Scale Machine Learning
    Guo, Tian-De
    Liu, Yan
    Han, Cong-Ying
    JOURNAL OF THE OPERATIONS RESEARCH SOCIETY OF CHINA, 2023, 11 (02) : 245 - 275
  • [30] An Overview of Stochastic Quasi-Newton Methods for Large-Scale Machine Learning
    Tian-De Guo
    Yan Liu
    Cong-Ying Han
    Journal of the Operations Research Society of China, 2023, 11 : 245 - 275