A parallel and distributed stochastic gradient descent implementation using commodity clusters

被引:0
|
作者
Robert K. L. Kennedy
Taghi M. Khoshgoftaar
Flavio Villanustre
Timothy Humphrey
机构
[1] Florida Atlantic University,
[2] LexisNexis Business Information Solutions,undefined
来源
关键词
Parallel stochastic gradient descent; Parallel and distributed processing; Deep learning; Big data; Neural network; Cluster computer; HPCC systems;
D O I
暂无
中图分类号
学科分类号
摘要
Deep Learning is an increasingly important subdomain of artificial intelligence, which benefits from training on Big Data. The size and complexity of the model combined with the size of the training dataset makes the training process very computationally and temporally expensive. Accelerating the training process of Deep Learning using cluster computers faces many challenges ranging from distributed optimizers to the large communication overhead specific to systems with off the shelf networking components. In this paper, we present a novel distributed and parallel implementation of stochastic gradient descent (SGD) on a distributed cluster of commodity computers. We use high-performance computing cluster (HPCC) systems as the underlying cluster environment for the implementation. We overview how the HPCC systems platform provides the environment for distributed and parallel Deep Learning, how it provides a facility to work with third party open source libraries such as TensorFlow, and detail our use of third-party libraries and HPCC functionality for implementation. We provide experimental results that validate our work and show that our implementation can scale with respect to both dataset size and the number of compute nodes in the cluster.
引用
收藏
相关论文
共 50 条
  • [21] Digital image correlation using stochastic parallel-gradient-descent algorithm
    Long, X.
    Fu, S.
    Qi, Z.
    Yang, X.
    Yu, Qifeng
    Proceedings of the Society for Experimental Mechanics, Inc., 2013, 53 (01): : 571 - 578
  • [22] Communication-Censored Distributed Stochastic Gradient Descent
    Li, Weiyu
    Wu, Zhaoxian
    Chen, Tianyi
    Li, Liping
    Ling, Qing
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2022, 33 (11) : 6831 - 6843
  • [23] Adaptive wavefront correction using a VLSI implementation of the parallel gradient descent algorithm
    Carhart, GW
    Vorontsov, MA
    Cohen, M
    Cauwenberghs, G
    Edwards, RT
    HIGH-RESOLUTION WAVEFRONT CONTROL: METHODS, DEVICES, AND APPLICATIONS, 1999, 3760 : 61 - 66
  • [24] Phase locking of sixteen laser beams using stochastic parallel gradient descent algorithm
    Zhou P.
    Wang X.
    Ma Y.
    Ma H.
    Xu X.
    Liu Z.
    Zhongguo Jiguang/Chinese Journal of Lasers, 2010, 37 (02): : 367 - 369
  • [25] Modal Analysis of Fiber Laser Beam by Using Stochastic Parallel Gradient Descent Algorithm
    Huang, Liangjin
    Lu, Haibin
    Zhou, Pu
    Leng, Jinyong
    Guo, Shaofeng
    Cheng, Xiang'ai
    IEEE PHOTONICS TECHNOLOGY LETTERS, 2015, 27 (21) : 2280 - 2283
  • [26] Wavefront error correction with stochastic parallel gradient descent algorithm
    Liu Jiaguo
    Li Lin
    Hu Xinqi
    Yu Xin
    Zhao Lei
    OPTICAL DESIGN AND TESTING III, PTS 1 AND 2, 2008, 6834
  • [27] Weighted Aggregating Stochastic Gradient Descent for Parallel Deep Learning
    Guo, Pengzhan
    Ye, Zeyang
    Xiao, Keli
    Zhu, Wei
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2022, 34 (10) : 5037 - 5050
  • [28] Test of the stochastic parallel gradient descent algorithm in laboratory experiments
    Banakh V.A.
    Larichev A.V.
    Razenkov I.A.
    Shesternin A.N.
    Atmospheric and Oceanic Optics, 2013, 26 (4) : 337 - 344
  • [29] Two-beam coherent addition using the stochastic parallel gradient descent algorithm
    Yang, Yuchuan
    Tan, Bitao
    Lizi, Chen
    Zhang, Junwei
    Wang, Xiao
    OPTICAL ENGINEERING, 2013, 52 (01)
  • [30] High Performance Parallel Stochastic Gradient Descent in Shared Memory
    Sallinen, Scott
    Satish, Nadathur
    Smelyanskiy, Mikhail
    Sury, Samantika S.
    Re, Christopher
    2016 IEEE 30TH INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM (IPDPS 2016), 2016, : 873 - 882