GPU in-memory processing using Spark for iterative computation

被引:17
|
作者
Hong, Sumin [1 ]
Choi, Woohyuk [1 ]
Jeong, Won-Ki [1 ]
机构
[1] Ulsan Natl Inst Sci & Technol, Sch Elect & Comp Engn, Ulsan, South Korea
基金
新加坡国家研究基金会;
关键词
Spark; MapReduce; GPU; In-memory Computing; FRAMEWORK;
D O I
10.1109/CCGRID.2017.41
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Due to its simplicity and scalability, MapReduce has become a de facto standard computing model for big data processing. Since the original MapReduce model was only appropriate for embarrassingly parallel batch processing, many follow-up studies have focused on improving the efficiency and performance of the model. Spark follows one of these recent trends by providing in-memory processing capability to reduce slow disk I/O for iterative computing tasks. However, the acceleration of Spark's in-memory processing using graphics processing units (GPUs) is challenging due to its deep memory hierarchy and host-to-GPU communication overhead. In this paper, we introduce a novel GPU-accelerated MapReduce framework that extends Spark's in-memory processing so that iterative computing is performed only in the GPU memory. Having discovered that the main bottleneck in the current Spark system for GPU computing is data communication on a Java virtual machine, we propose a modification of the current Spark implementation to bypass expensive data management for iterative task offloading to GPUs. We also propose a novel GPU in-memory processing and caching framework that minimizes host-to-GPU communication via lazy evaluation and reuses GPU memory over multiple mapper executions. The proposed system employs message-passing interface (MPI)-based data synchronization for inter-worker communication so that more complicated iterative computing tasks, such as iterative numerical solvers, can be efficiently handled. We demonstrate the performance of our system in terms of several iterative computing tasks in big data processing applications, including machine learning and scientific computing. We achieved up to 50 times speed up over conventional Spark and about 10 times speed up over GPU-accelerated Spark.
引用
收藏
页码:31 / 41
页数:11
相关论文
共 50 条
  • [21] Fast In-Memory Transaction Processing Using RDMA and HTM
    Chen, Haibo
    Chen, Rong
    Wei, Xingda
    Shi, Jiaxin
    Chen, Yanzhe
    Wang, Zhaoguo
    Zang, Binyu
    Guan, Haibing
    [J]. ACM TRANSACTIONS ON COMPUTER SYSTEMS, 2017, 35 (01):
  • [22] Processing Data Where It Makes Sense in Modern Computing Systems: Enabling In-Memory Computation
    Mutlu, Onur
    [J]. 2018 7TH MEDITERRANEAN CONFERENCE ON EMBEDDED COMPUTING (MECO), 2018, : 8 - 9
  • [23] Enabling In-Memory Computation of Binary BLAS using ReRAM Crossbar Arrays
    Bhattacharjee, Debjyoti
    Merchant, Farhad
    Chattopadhyay, Anupam
    [J]. 2016 IFIP/IEEE INTERNATIONAL CONFERENCE ON VERY LARGE SCALE INTEGRATION (VLSI-SOC), 2016,
  • [24] Processing Data Where It Makes Sense in Modern Computing Systems: Enabling In-Memory Computation
    Mutlu, Onur
    [J]. GLSVLSI '19 - PROCEEDINGS OF THE 2019 ON GREAT LAKES SYMPOSIUM ON VLSI, 2019, : 5 - 6
  • [25] Revamped Market-Basket Analysis Using In-Memory Computation Framework
    Thanmayee
    Prasad, H. R. Manjunath
    [J]. PROCEEDINGS OF 2017 11TH INTERNATIONAL CONFERENCE ON INTELLIGENT SYSTEMS AND CONTROL (ISCO 2017), 2017, : 65 - 70
  • [26] Memristive Memory Processing Unit (MPU) Controller for In-Memory Processing
    Ben Hur, Rotem
    Kvatinsky, Shahar
    [J]. 2016 IEEE INTERNATIONAL CONFERENCE ON THE SCIENCE OF ELECTRICAL ENGINEERING (ICSEE), 2016,
  • [27] Ignite-GPU: a GPU-enabled in-memory computing architecture on clusters
    Sojoodi, Amir Hossein
    Salimi Beni, Majid
    Khunjush, Farshad
    [J]. JOURNAL OF SUPERCOMPUTING, 2021, 77 (03): : 3165 - 3192
  • [28] Ignite-GPU: a GPU-enabled in-memory computing architecture on clusters
    Amir Hossein Sojoodi
    Majid Salimi Beni
    Farshad Khunjush
    [J]. The Journal of Supercomputing, 2021, 77 : 3165 - 3192
  • [29] MPIM: Multi-Purpose In-Memory Processing Using Configurable Resistive Memory
    Imani, Mohsen
    Kim, Yeseong
    Rosing, Tajana
    [J]. 2017 22ND ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2017, : 757 - 763
  • [30] Accelerating in-memory transaction processing using general purpose graphics processing units
    Gao, Lan
    Xu, Yunlong
    Wang, Rui
    Yang, Hailong
    Luan, Zhongzhi
    Qian, Depei
    [J]. FUTURE GENERATION COMPUTER SYSTEMS-THE INTERNATIONAL JOURNAL OF ESCIENCE, 2019, 97 : 836 - 848