A parallel and distributed stochastic gradient descent implementation using commodity clusters

被引:0
|
作者
Robert K. L. Kennedy
Taghi M. Khoshgoftaar
Flavio Villanustre
Timothy Humphrey
机构
[1] Florida Atlantic University,
[2] LexisNexis Business Information Solutions,undefined
来源
关键词
Parallel stochastic gradient descent; Parallel and distributed processing; Deep learning; Big data; Neural network; Cluster computer; HPCC systems;
D O I
暂无
中图分类号
学科分类号
摘要
Deep Learning is an increasingly important subdomain of artificial intelligence, which benefits from training on Big Data. The size and complexity of the model combined with the size of the training dataset makes the training process very computationally and temporally expensive. Accelerating the training process of Deep Learning using cluster computers faces many challenges ranging from distributed optimizers to the large communication overhead specific to systems with off the shelf networking components. In this paper, we present a novel distributed and parallel implementation of stochastic gradient descent (SGD) on a distributed cluster of commodity computers. We use high-performance computing cluster (HPCC) systems as the underlying cluster environment for the implementation. We overview how the HPCC systems platform provides the environment for distributed and parallel Deep Learning, how it provides a facility to work with third party open source libraries such as TensorFlow, and detail our use of third-party libraries and HPCC functionality for implementation. We provide experimental results that validate our work and show that our implementation can scale with respect to both dataset size and the number of compute nodes in the cluster.
引用
收藏
相关论文
共 50 条
  • [1] A parallel and distributed stochastic gradient descent implementation using commodity clusters
    Kennedy, Robert K. L.
    Khoshgoftaar, Taghi M.
    Villanustre, Flavio
    Humphrey, Timothy
    JOURNAL OF BIG DATA, 2019, 6 (01)
  • [2] Parallel Implementation on FPGA of Support Vector Machines Using Stochastic Gradient Descent
    Lopes, Felipe E.
    Ferreira, Joao Canas
    Fernandes, Marcelo A. C.
    ELECTRONICS, 2019, 8 (06)
  • [3] Adaptive wavefront control with asynchronous stochastic parallel gradient descent clusters
    Vorontsov, Mikhail A.
    Carhart, Gary W.
    JOURNAL OF THE OPTICAL SOCIETY OF AMERICA A-OPTICS IMAGE SCIENCE AND VISION, 2006, 23 (10) : 2613 - 2622
  • [4] Distributed Stochastic Gradient Descent Using LDGM Codes
    Horii, Shunsuke
    Yoshida, Takahiro
    Kobayashi, Manabu
    Matsushima, Toshiyasu
    2019 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2019, : 1417 - 1421
  • [5] Implementation of Stochastic Parallel Gradient Descent Algorithm for Coherent Beam Combining
    Linslal, C. L.
    Sooraj, M. S.
    Padmanabhan, A.
    Venkitesh, D.
    Srinivasan, B.
    HIGH-POWER LASERS AND APPLICATIONS IX, 2018, 10811
  • [6] Bayesian Distributed Stochastic Gradient Descent
    Teng, Michael
    Wood, Frank
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [7] Implementation and performance of stochastic parallel gradient descent algorithm for atmospheric turbulence compensation
    Finney, Greg A.
    Persons, Christopher
    Henning, Stephan
    Hazen, Jessie
    Whitley, Daniel
    LASER RADAR TECHNOLOGY AND APPLICATIONS XIX; AND ATMOSPHERIC PROPAGATION XI, 2014, 9080
  • [8] CONTROLLING STOCHASTIC GRADIENT DESCENT USING STOCHASTIC APPROXIMATION FOR ROBUST DISTRIBUTED OPTIMIZATION
    Jain, Adit
    Krishnamurthy, Vikram
    NUMERICAL ALGEBRA CONTROL AND OPTIMIZATION, 2024,
  • [9] Asynchronous Decentralized Parallel Stochastic Gradient Descent
    Lian, Xiangru
    Zhang, Wei
    Zhang, Ce
    Liu, Ji
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [10] The Impact of Synchronization in Parallel Stochastic Gradient Descent
    Backstrom, Karl
    Papatriantafilou, Marina
    Tsigas, Philippas
    DISTRIBUTED COMPUTING AND INTELLIGENT TECHNOLOGY, ICDCIT 2022, 2022, 13145 : 60 - 75