FTSGD: An Adaptive Stochastic Gradient Descent Algorithm for Spark MLlib

被引:2
|
作者
Zhang, Hong [1 ]
Liu, Zixia [1 ]
Huang, Hai [2 ]
Wang, Liqiang [1 ]
机构
[1] Univ Cent Florida, Dept Comp Sci, Orlando, FL 32816 USA
[2] IBM TJ Watson Res Ctr, Yorktown Hts, NY USA
关键词
Spark; MLlib; Asynchronous Stochastic Gradient Decent; Adaptive Iterative Learning;
D O I
10.1109/DASC/PiCom/DataCom/CyberSciTec.2018.00-22
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The proliferation of massive datasets and the surge of interests in big data analytics have popularized a number of novel distributed data processing platforms such as Hadoop and Spark. Their large and growing ecosystems of libraries enable even novice to take advantage of the latest data analytics and machine learning algorithms. However, time-consuming data synchronization and communications in iterative algorithms on large-scale distributed platforms can lead to significant performance inefficiency. MLlib is Spark's scalable library consisting of common machine learning algorithms, many of which employ Stochastic Gradient Descent (SGD) to find minima or maxima by iterations. However, the convergence can be very slow if gradient data are synchronized on each iteration. In this work, we optimize the current implementation of SGD in Spark's MLlib by reusing data partition for multiple times within a single iteration to find better candidate weights in a more efficient way. Whether using multiple local iterations within each partition is dynamically decided by the 68-95-99.7 rule. We also design a variant of momentum algorithm to optimize step size in every iteration. This method uses a new adaptive rule that decreases the step size whenever neighboring gradients show differing directions of significance. Experiments show that our adaptive algorithm is more efficient and can be 7 times faster compared to the original MLlib's SGD.
引用
收藏
页码:828 / 835
页数:8
相关论文
共 50 条
  • [21] Adaptive Stochastic Gradient Descent Optimisation for Image Registration
    Stefan Klein
    Josien P. W. Pluim
    Marius Staring
    Max A. Viergever
    International Journal of Computer Vision, 2009, 81
  • [22] Adaptive Stochastic Gradient Descent Optimisation for Image Registration
    Klein, Stefan
    Pluim, Josien P. W.
    Staring, Marius
    Viergever, Max A.
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2009, 81 (03) : 227 - 239
  • [23] ADINE: An Adaptive Momentum Method for Stochastic Gradient Descent
    Srinivasan, Vishwak
    Sankar, Adepu Ravi
    Balasubramanian, Vineeth N.
    PROCEEDINGS OF THE ACM INDIA JOINT INTERNATIONAL CONFERENCE ON DATA SCIENCE AND MANAGEMENT OF DATA (CODS-COMAD'18), 2018, : 249 - 256
  • [24] A new stochastic gradient descent possibilistic clustering algorithm
    Koutsimpela, Angeliki
    Koutroumbas, Konstantinos D.
    AI COMMUNICATIONS, 2022, 35 (02) : 47 - 64
  • [25] Fast Convergence Stochastic Parallel Gradient Descent Algorithm
    Hu Dongting
    Shen Wen
    Ma Wenchao
    Liu Xinyu
    Su Zhouping
    Zhu Huaxin
    Zhang Xiumei
    Que Lizhi
    Zhu Zhuowei
    Zhang Yixin
    Chen Guoqing
    Hu Lifa
    LASER & OPTOELECTRONICS PROGRESS, 2019, 56 (12)
  • [26] Guided Stochastic Gradient Descent Algorithm for inconsistent datasets
    Sharma, Anuraganand
    APPLIED SOFT COMPUTING, 2018, 73 : 1068 - 1080
  • [27] Stochastic Approximate Gradient Descent via the Langevin Algorithm
    Qiu, Yixuan
    Wang, Xiao
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 5428 - 5435
  • [28] A stochastic gradient descent algorithm for structural risk minimisation
    Ratsaby, J
    ALGORITHMIC LEARNING THEORY, PROCEEDINGS, 2003, 2842 : 205 - 220
  • [29] The Improved Stochastic Fractional Order Gradient Descent Algorithm
    Yang, Yang
    Mo, Lipo
    Hu, Yusen
    Long, Fei
    FRACTAL AND FRACTIONAL, 2023, 7 (08)
  • [30] Convergence behavior of diffusion stochastic gradient descent algorithm
    Barani, Fatemeh
    Savadi, Abdorreza
    Yazdi, Hadi Sadoghi
    SIGNAL PROCESSING, 2021, 183