RECENT TRENDS IN STOCHASTIC GRADIENT DESCENT FOR MACHINE LEARNING AND BIG DATA

被引:0
|
作者
Newton, David [1 ]
Pasupathy, Raghu [1 ]
Yousefian, Farzad [2 ]
机构
[1] Purdue Univ, Dept Stat, W Lafayette, IN 47906 USA
[2] Oklahoma State Univ, Dept Ind Engn & Management, Stillwater, OK 74078 USA
关键词
SUBGRADIENT METHODS; APPROXIMATION;
D O I
暂无
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Stochastic Gradient Descent (SGD), also known as stochastic approximation, refers to certain simple iterative structures used for solving stochastic optimization and root finding problems. The identifying feature of SGD is that, much like in gradient descent for deterministic optimization, each successive iterate in the recursion is determined by adding an appropriately scaled gradient estimate to the prior iterate. Owing to several factors, SGD has become the leading method to solve optimization problems arising within large-scale machine learning and "big data" contexts such as classification and regression. This tutorial covers the basics of SGD with an emphasis on modern developments. The tutorial starts with examples where SGD is applicable, and then details important flavors of SGD and reported complexity calculations.
引用
收藏
页码:366 / 380
页数:15
相关论文
共 50 条
  • [1] From big data to smart data: a sample gradient descent approach for machine learning
    Ganie, Aadil Gani
    Dadvandipour, Samad
    [J]. JOURNAL OF BIG DATA, 2023, 10 (01)
  • [2] From big data to smart data: a sample gradient descent approach for machine learning
    Aadil Gani Ganie
    Samad Dadvandipour
    [J]. Journal of Big Data, 10
  • [3] Recent Advances in Stochastic Gradient Descent in Deep Learning
    Tian, Yingjie
    Zhang, Yuqi
    Zhang, Haibin
    [J]. MATHEMATICS, 2023, 11 (03)
  • [4] Stochastic Gradient Descent and Its Variants in Machine Learning
    Netrapalli, Praneeth
    [J]. JOURNAL OF THE INDIAN INSTITUTE OF SCIENCE, 2019, 99 (02) : 201 - 213
  • [5] Stochastic Gradient Descent and Its Variants in Machine Learning
    Praneeth Netrapalli
    [J]. Journal of the Indian Institute of Science, 2019, 99 : 201 - 213
  • [6] Large-Scale Machine Learning with Stochastic Gradient Descent
    Bottou, Leon
    [J]. COMPSTAT'2010: 19TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL STATISTICS, 2010, : 177 - 186
  • [7] Distributed Byzantine Tolerant Stochastic Gradient Descent in the Era of Big Data
    Jin, Richeng
    He, Xiaofan
    Dai, Huaiyu
    [J]. ICC 2019 - 2019 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS (ICC), 2019,
  • [8] In-Database Machine Learning with CorgiPile: Stochastic Gradient Descent without Full Data Shuffle
    Xu, Lijie
    Qiu, Shuang
    Yuan, Binhang
    Jiang, Jiawei
    Renggli, Cedric
    Gan, Shaoduo
    Kara, Kaan
    Li, Guoliang
    Liu, Ji
    Wu, Wentao
    Ye, Jieping
    Zhang, Ce
    [J]. PROCEEDINGS OF THE 2022 INTERNATIONAL CONFERENCE ON MANAGEMENT OF DATA (SIGMOD '22), 2022, : 1286 - 1300
  • [9] Coded Decentralized Learning With Gradient Descent for Big Data Analytics
    Yue, Jing
    Xiao, Ming
    [J]. IEEE COMMUNICATIONS LETTERS, 2020, 24 (02) : 362 - 366
  • [10] Big data and machine learning driven bioprocessing-Recent trends and critical analysis
    Yang, Chao-Tung
    Kristiani, Endah
    Leong, Yoong Kit
    Chang, Jo-Shu
    [J]. BIORESOURCE TECHNOLOGY, 2023, 372