Iterative Sparse Matrix-Vector Multiplication on In-Memory Cluster Computing Accelerated by GPUs for Big Data

被引:0
|
作者
Peng, Jiwu [1 ,2 ]
Xiao, Zheng [1 ,2 ]
Chen, Cen [1 ,2 ]
Yang, Wangdong [1 ,2 ]
机构
[1] Hunan Univ, Coll Informat Sci & Engn, Changsha 410082, Hunan, Peoples R China
[2] Natl Supercomp Ctr Changsha, Changsha 410082, Hunan, Peoples R China
关键词
Iterative SpMV; Flink; GPU; In-memory Computing; BigData;
D O I
暂无
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Iterative SpMV (ISpMV) is a key operation in many graph-based data mining algorithms and machine learning algorithms. Along with the development of big data, the matrices can be so large, perhaps billion-scale, that the SpMV can not be implemented in a single computer. Therefore, it is a challenging issue to implement and optimize SpMV for large-scale data sets. In this paper, we used an in-memory heterogeneous CPU-GPU cluster computing platforms (IMHCPs) to efficiently solve billion-scale SpMV problem. A dedicated and efficient hierarchy partitioning strategy for sparse matrices and the vector is proposed. The partitioning strategy contains partitioning sparse matrices among workers in the cluster and among GPUs in one worker. More, the performance of the IMHCPs-based SpMV is evaluated from the aspects of computation efficiency and scalability.
引用
收藏
页码:1454 / 1460
页数:7
相关论文
共 50 条
  • [21] Automatic Tuning of Sparse Matrix-Vector Multiplication for CRS format on GPUs
    Yoshizawa, Hiroki
    Takahashi, Daisuke
    15TH IEEE INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND ENGINEERING (CSE 2012) / 10TH IEEE/IFIP INTERNATIONAL CONFERENCE ON EMBEDDED AND UBIQUITOUS COMPUTING (EUC 2012), 2012, : 130 - 136
  • [22] A Memory Transaction Model for Sparse Matrix-Vector Multiplications on GPUs
    Keklikian, Thalie
    Langlois, J. M. Pierre
    Savaria, Yvon
    2014 IEEE 12TH INTERNATIONAL NEW CIRCUITS AND SYSTEMS CONFERENCE (NEWCAS), 2014, : 309 - 312
  • [23] Communication Optimization of Iterative Sparse Matrix-Vector Multiply on GPUs and FPGAs
    Rafique, Abid
    Constantinides, George A.
    Kapre, Nachiket
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2015, 26 (01) : 24 - 34
  • [24] Scalable In-Memory Computing Architectures for Sparse Matrix Multiplication
    Kendall, Jack D.
    Conklin, Alexander A.
    Pantone, Ross
    Nino, Juan C.
    Kumar, Suhas
    2022 INTERNATIONAL ELECTRON DEVICES MEETING, IEDM, 2022,
  • [25] An architecture-aware technique for optimizing sparse matrix-vector multiplication on GPUs
    Maggioni, Marco
    Berger-Wolf, Tanya
    2013 INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE, 2013, 18 : 329 - 338
  • [26] Efficient Sparse Matrix-Vector Multiplication on GPUs using the CSR Storage Format
    Greathouse, Joseph L.
    Daga, Mayank
    SC14: INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, 2014, : 769 - 780
  • [27] A Performance Modeling and Optimization Analysis Tool for Sparse Matrix-Vector Multiplication on GPUs
    Guo, Ping
    Wang, Liqiang
    Chen, Po
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2014, 25 (05) : 1112 - 1123
  • [28] Programming Weights to Analog In-Memory Computing Cores by Direct Minimization of the Matrix-Vector Multiplication Error
    Buechel, Julian
    Vasilopoulos, Athanasios
    Kersting, Benedikt
    Lammie, Corey
    Brew, Kevin
    Philip, Timothy
    Saulnier, Nicole
    Narayanan, Vijay
    Le Gallo, Manuel
    Sebastian, Abu
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2023, 13 (04) : 1052 - 1061
  • [29] Sparse Matrix-Vector Multiplication on GPGPUs
    Filippone, Salvatore
    Cardellini, Valeria
    Barbieri, Davide
    Fanfarillo, Alessandro
    ACM TRANSACTIONS ON MATHEMATICAL SOFTWARE, 2017, 43 (04):
  • [30] Optimization of Sparse Matrix-Vector Multiplication for CRS Format on NVIDIA Kepler Architecture GPUs
    Mukunoki, Daichi
    Takahashi, Daisuke
    COMPUTATIONAL SCIENCE AND ITS APPLICATIONS - ICCSA 2013, PT V, 2013, 7975 : 211 - 223