RECENT TRENDS IN STOCHASTIC GRADIENT DESCENT FOR MACHINE LEARNING AND BIG DATA

被引:0
|
作者
Newton, David [1 ]
Pasupathy, Raghu [1 ]
Yousefian, Farzad [2 ]
机构
[1] Purdue Univ, Dept Stat, W Lafayette, IN 47906 USA
[2] Oklahoma State Univ, Dept Ind Engn & Management, Stillwater, OK 74078 USA
关键词
SUBGRADIENT METHODS; APPROXIMATION;
D O I
暂无
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Stochastic Gradient Descent (SGD), also known as stochastic approximation, refers to certain simple iterative structures used for solving stochastic optimization and root finding problems. The identifying feature of SGD is that, much like in gradient descent for deterministic optimization, each successive iterate in the recursion is determined by adding an appropriately scaled gradient estimate to the prior iterate. Owing to several factors, SGD has become the leading method to solve optimization problems arising within large-scale machine learning and "big data" contexts such as classification and regression. This tutorial covers the basics of SGD with an emphasis on modern developments. The tutorial starts with examples where SGD is applicable, and then details important flavors of SGD and reported complexity calculations.
引用
收藏
页码:366 / 380
页数:15
相关论文
共 50 条
  • [21] Stochastic Gradient Descent on Separable Data: Exact Convergence with a Fixed Learning Rate
    Nacson, Mor Shpigel
    Srebro, Nathan
    Soudry, Daniel
    [J]. 22ND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 89, 2019, 89
  • [22] Trends of Evolutionary Machine Learning to Address Big Data Mining
    Ben Hamida, Sana
    Benjelloun, Ghita
    Hmida, Hmida
    [J]. INFORMATION AND KNOWLEDGE SYSTEMS: DIGITAL TECHNOLOGIES, ARTIFICIAL INTELLIGENCE AND DECISION MAKING, ICIKS 2021, 2021, 425 : 85 - 99
  • [23] Asymptotic Network Independence in Distributed Stochastic Optimization for Machine Learning: Examining Distributed and Centralized Stochastic Gradient Descent
    Pu, Shi
    Olshevsky, Alex
    Paschalidis, Ioannis Ch.
    [J]. IEEE SIGNAL PROCESSING MAGAZINE, 2020, 37 (03) : 114 - 122
  • [24] Machine-learning topology optimization with stochastic gradient descent optimizer for heat conduction problems
    Hua, Yuchao
    Luo, Lingai
    Le Corre, Steven
    Fan, Yilin
    [J]. INTERNATIONAL JOURNAL OF HEAT AND MASS TRANSFER, 2024, 223
  • [25] Stochastic Gradient Descent with Noise of Machine Learning Type Part II: Continuous Time Analysis
    Wojtowytsch S.
    [J]. Journal of Nonlinear Science, 2024, 34 (1)
  • [26] Stochastic Gradient Descent with Noise of Machine Learning Type Part I: Discrete Time Analysis
    Wojtowytsch, Stephan
    [J]. JOURNAL OF NONLINEAR SCIENCE, 2023, 33 (03)
  • [27] Towards Learning Stochastic Population Models by Gradient Descent
    Kreikemeyer, Justin N.
    Andelfinger, Philipp
    Uhrmacher, Adelinde M.
    [J]. PROCEEDINGS OF THE 38TH ACM SIGSIM INTERNATIONAL CONFERENCE ON PRINCIPLES OF ADVANCED DISCRETE SIMULATION, ACM SIGSIM-PADS 2024, 2024, : 88 - 92
  • [28] Stochastic Gradient Descent with Noise of Machine Learning Type Part I: Discrete Time Analysis
    Stephan Wojtowytsch
    [J]. Journal of Nonlinear Science, 2023, 33
  • [29] Stochastic Gradient Descent with Polyak's Learning Rate
    Prazeres, Mariana
    Oberman, Adam M.
    [J]. JOURNAL OF SCIENTIFIC COMPUTING, 2021, 89 (01)
  • [30] From Gradient Flow on Population Loss to Learning with Stochastic Gradient Descent
    Sekhari, Ayush
    Kale, Satyen
    Lee, Jason D.
    De Sa, Chris
    Sridharan, Karthik
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,