The Frontier of SGD and Its Variants in Machine Learning

被引:4
|
作者
Du, Juan [1 ]
机构
[1] New Res & Dev Ctr Hisense, Qingdao 266071, Peoples R China
关键词
D O I
10.1088/1742-6596/1229/1/012046
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
A Numerical optimization is a classical field in operation research and computer science, which has been widely used in areas such as physics and economics. Although optimization algorithms have achieved great success for plenty of applications, handling the big data in the best fashion possible is a very inspiring and demanding challenge in the artificial intelligence era. Stochastic gradient descent (SGD) is pretty simple but surprisingly, highly effective in machine learning models, such as support vector machine (SVM) and deep neural network (DNN). Theoretically, the performance of SGD for convex optimization is well understood. But, for the non-convex setting, which is very common for the machine learning problems, to obtain the theoretical guarantee for SGD and its variants is still a standing problem. In the paper, we do a survey about the SGD and its variants such as Momentum, ADAM and SVRG, differentiate their algorithms and applications and present some recent breakthrough and open problems.
引用
收藏
页数:8
相关论文
共 50 条
  • [1] The Frontier of SGD and its Variants in Machine Learning
    Du, Juan
    [J]. 2017 2ND INTERNATIONAL CONFERENCE ON MECHATRONICS AND INFORMATION TECHNOLOGY (ICMIT 2017), 2017, : 360 - 366
  • [2] Asynchronous Byzantine Machine Learning (the case of SGD)
    Damaskinos, Georgios
    El Mhamdi, El Mandi
    Guerraoui, Rachid
    Patra, Rhicheek
    Taziki, Mahsa
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [3] Stochastic Gradient Descent and Its Variants in Machine Learning
    Praneeth Netrapalli
    [J]. Journal of the Indian Institute of Science, 2019, 99 : 201 - 213
  • [4] Stochastic Gradient Descent and Its Variants in Machine Learning
    Netrapalli, Praneeth
    [J]. JOURNAL OF THE INDIAN INSTITUTE OF SCIENCE, 2019, 99 (02) : 201 - 213
  • [5] CL-SGD: Efficient Communication by Clustering and Local-SGD for Distributed Machine Learning
    Chen, Haosen
    Liu, Wai-xi
    Tan, Miaoquan
    Luo, Junming
    [J]. ICC 2023-IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS, 2023, : 4409 - 4415
  • [6] A machine learning efficient frontier
    Clark, Brian
    Feinstein, Zachary
    Simaan, Majeed
    [J]. OPERATIONS RESEARCH LETTERS, 2020, 48 (05) : 630 - 634
  • [7] Investigating translated Chinese and its variants using machine learning
    Hu, Hai
    Kubler, Sandra
    [J]. NATURAL LANGUAGE ENGINEERING, 2021, 27 (03) : 339 - 372
  • [8] CE-SGD: Communication-Efficient Distributed Machine Learning
    Tao, Zeyi
    Xia, Qi
    Li, Qun
    Cheng, Songqing
    [J]. 2021 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2021,
  • [9] Machine Learning Frontier Orbital Energies of Nanodiamonds
    Kirschbaum, Thorren
    Dzubiella, Joachim
    Bande, Annika
    Noe, Frank
    von Seggern, Boerries
    [J]. JOURNAL OF CHEMICAL THEORY AND COMPUTATION, 2023, 19 (14) : 4461 - 4473
  • [10] Image Reconstruction Is a New Frontier of Machine Learning
    Wang, Ge
    Ye, Jong Chu
    Mueller, Klaus
    Fessler, Jeffrey A.
    [J]. IEEE TRANSACTIONS ON MEDICAL IMAGING, 2018, 37 (06) : 1289 - 1296