Performance evaluation of the sparse matrix-vector multiplication on modern architectures

被引:58
|
作者
Goumas, Georgios [1 ]
Kourtis, Kornilios [1 ]
Anastopoulos, Nikos [1 ]
Karakasis, Vasileios [1 ]
Koziris, Nectarios [1 ]
机构
[1] Natl Tech Univ Athens, Comp Syst Lab, Sch Elect & Comp Engn, Zografos 15780, Greece
来源
JOURNAL OF SUPERCOMPUTING | 2009年 / 50卷 / 01期
关键词
Sparse matrix-vector multiplication; Multicore architectures; Scientific applications; Performance evaluation;
D O I
10.1007/s11227-008-0251-8
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we revisit the performance issues of the widely used sparse matrix-vector multiplication (SpMxV) kernel on modern microarchitectures. Previous scientific work reports a number of different factors that may significantly reduce performance. However, the interaction of these factors with the underlying architectural characteristics is not clearly understood, a fact that may lead to misguided, and thus unsuccessful attempts for optimization. In order to gain an insight into the details of SpMxV performance, we conduct a suite of experiments on a rich set of matrices for three different commodity hardware platforms. In addition, we investigate the parallel version of the kernel and report on the corresponding performance results and their relation to each architecture's specific multithreaded configuration. Based on our experiments, we extract useful conclusions that can serve as guidelines for the optimization process of both single and multithreaded versions of the kernel.
引用
收藏
页码:36 / 77
页数:42
相关论文
共 50 条
  • [41] Sparse Matrix-Vector Multiplication on a Reconfigurable Supercomputer with Application
    Dubois, David
    Dubois, Andrew
    Boorman, Thomas
    Connor, Carolyn
    Poole, Steve
    ACM TRANSACTIONS ON RECONFIGURABLE TECHNOLOGY AND SYSTEMS, 2010, 3 (01)
  • [42] Sparse Binary Matrix-Vector Multiplication on Neuromorphic Computers
    Schuman, Catherine D.
    Kay, Bill
    Date, Prasanna
    Kannan, Ramakrishnan
    Sao, Piyush
    Potok, Thomas E.
    2021 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS (IPDPSW), 2021, : 308 - 311
  • [43] Optimization techniques for sparse matrix-vector multiplication on GPUs
    Maggioni, Marco
    Berger-Wolf, Tanya
    JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING, 2016, 93-94 : 66 - 86
  • [44] Parallel Sparse Matrix-Vector Multiplication Using Accelerators
    Maeda, Hiroshi
    Takahashi, Daisuke
    COMPUTATIONAL SCIENCE AND ITS APPLICATIONS - ICCSA 2016, PT II, 2016, 9787 : 3 - 18
  • [45] Adaptive diagonal sparse matrix-vector multiplication on GPU
    Gao, Jiaquan
    Xia, Yifei
    Yin, Renjie
    He, Guixia
    JOURNAL OF PARALLEL AND DISTRIBUTED COMPUTING, 2021, 157 : 287 - 302
  • [46] Acceleration of Sparse Matrix-Vector Multiplication by Region Traversal
    Simecek, I.
    ACTA POLYTECHNICA, 2008, 48 (04) : 8 - 15
  • [47] IMAGE EDITING BASED ON SPARSE MATRIX-VECTOR MULTIPLICATION
    Wang, Ying
    Yan, Hongping
    Pan, Chunhong
    Xiang, Shiming
    2011 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2011, : 1317 - 1320
  • [48] Processor-efficient sparse matrix-vector multiplication
    Heath, LS
    Ribbens, CJ
    Pemmaraju, SV
    COMPUTERS & MATHEMATICS WITH APPLICATIONS, 2004, 48 (3-4) : 589 - 608
  • [49] A Comprehensive Performance Model of Sparse Matrix-Vector Multiplication to Guide Kernel Optimization
    Xia, Tian
    Fu, Gelin
    Li, Chenyang
    Luo, Zhongpei
    Zhang, Lucheng
    Chen, Ruiyang
    Zhao, Wenzhe
    Zheng, Nanning
    Ren, Pengju
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2023, 34 (02) : 519 - 534
  • [50] Performance modeling and optimization of sparse matrix-vector multiplication on NVIDIA CUDA platform
    Shiming Xu
    Wei Xue
    Hai Xiang Lin
    The Journal of Supercomputing, 2013, 63 : 710 - 721