GPU-Accelerated Parallel Hierarchical Extreme Learning Machine on Flink for Big Data

被引:81
|
作者
Chen, Cen [1 ,2 ]
Li, Kenli [1 ,2 ]
Ouyang, Aijia [1 ,2 ,3 ]
Tang, Zhuo [1 ,2 ]
Li, Keqin [1 ,2 ,4 ]
机构
[1] Hunan Univ, Coll Informat Sci & Engn, Changsha 410082, Hunan, Peoples R China
[2] Natl Supercomp Ctr, Changsha 410082, Hunan, Peoples R China
[3] Zunyi Normal Coll, Dept Informat Engn, Zunyi 563006, Peoples R China
[4] SUNY Coll New Paltz, Dept Comp Sci, New Paltz, NY 12561 USA
基金
中国国家自然科学基金;
关键词
Big data; deep learning (DL); Flink; GPGPU; hierarchical extreme learning machine (H-ELM); parallel; FEEDFORWARD NETWORKS; HIDDEN NODES; MAPREDUCE; APPROXIMATION; CLASSIFICATION; OPTIMIZATION; REGRESSION; ALGORITHM; SPMV;
D O I
10.1109/TSMC.2017.2690673
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The extreme learning machine (ELM) has become one of the most important and popular algorithms of machine learning, because of its extremely fast training speed, good generalization, and universal approximation/classification capability. The proposal of hierarchical ELM (H-ELM) extends ELM from single hidden layer feedforward networks to multilayer perceptron, greatly strengthening the applicability of ELM. Generally speaking, during training H-ELM, large-scale datasets (DSTs) are needed. Therefore, how to make use of H-ELM framework in processing big data is worth further exploration. This paper proposes a parallel H-ELM algorithm based on Flink, which is one of the in-memory cluster computing platforms, and graphics processing units (GPUs). Several optimizations are adopted to improve the performance, such as cache-based scheme, reasonable partitioning strategy, memory mapping scheme for mapping specific Java virtual machine objects to buffers. Most importantly, our proposed framework for utilizing GPUs to accelerate Flink for big data is general. This framework can be utilized to accelerate many other variants of ELM and other machine learning algorithms. To the best of our knowledge, it is the first kind of library, which combines in-memory cluster computing with GPUs to parallelize H-ELM. The experimental results have demonstrated that our proposed GPU-accelerated parallel H-ELM named as GPH-ELM can efficiently process large-scale DSTs with good performance of speedup and scalability, leveraging the computing power of both CPUs and GPUs in the cluster.
引用
收藏
页码:2740 / 2753
页数:14
相关论文
共 50 条
  • [1] GPU-accelerated and mixed norm regularized online extreme learning machine
    Polat, Onder
    Kayhan, Sema Koc
    [J]. CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2022, 34 (15):
  • [2] GPU-Accelerated Extreme Learning Machines for Imbalanced Data Streams with Concept Drift
    Krawczyk, Bartosz
    [J]. INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE 2016 (ICCS 2016), 2016, 80 : 1692 - 1701
  • [3] Interactive machine learning via a GPU-accelerated Toolkit
    Jiang, Biye
    Canny, John
    [J]. IUI'17: PROCEEDINGS OF THE 22ND INTERNATIONAL CONFERENCE ON INTELLIGENT USER INTERFACES, 2017, : 535 - 546
  • [4] A Survey of Distributed and Parallel Extreme Learning Machine for Big Data
    Wang, Zhiqiong
    Sui, Ling
    Xin, Junchang
    Qu, Luxuan
    Yao, Yudong
    [J]. IEEE ACCESS, 2020, 8 : 201247 - 201258
  • [5] GPU-accelerated approximate kernel method for quantum machine learning
    Browning, Nicholas J.
    Faber, Felix A.
    von Lilienfeld, O. Anatole
    [J]. JOURNAL OF CHEMICAL PHYSICS, 2022, 157 (21):
  • [6] Distributed parallel deep learning of Hierarchical Extreme Learning Machine for multimode quality prediction with big process data
    Yao, Le
    Ge, Zhiqiang
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2019, 81 : 450 - 465
  • [7] GPU-accelerated parallel optimization for sparse regularization
    Wang, Xingran
    Liu, Tianyi
    Minh Trinh-Hoang
    Pesavento, Marius
    [J]. 2020 IEEE 11TH SENSOR ARRAY AND MULTICHANNEL SIGNAL PROCESSING WORKSHOP (SAM), 2020,
  • [8] Tuned and GPU-Accelerated Parallel Data Mining from Comparable Corpora
    Wolk, Krzysztof
    Marasek, Krzysztof
    [J]. TEXT, SPEECH, AND DIALOGUE (TSD 2015), 2015, 9302 : 32 - 40
  • [9] GPU-accelerated parallel algorithms for linear rankSVM
    Jin, Jing
    Cai, Xianggao
    Lai, Guoming
    Lin, Xiaola
    [J]. JOURNAL OF SUPERCOMPUTING, 2015, 71 (11): : 4141 - 4171
  • [10] GPU-accelerated parallel algorithms for linear rankSVM
    Jing Jin
    Xianggao Cai
    Guoming Lai
    Xiaola Lin
    [J]. The Journal of Supercomputing, 2015, 71 : 4141 - 4171