Input-Aware Sparse Tensor Storage Format Selection for Optimizing MTTKRP

被引:5
|
作者
Sun, Qingxiao [1 ]
Liu, Yi [1 ]
Yang, Hailong [1 ]
Dun, Ming [2 ]
Luan, Zhongzhi [1 ]
Gan, Lin [3 ]
Yang, Guangwen [3 ]
Qian, Depei [1 ]
机构
[1] Beihang Univ, Sch Comp Sci & Engn, Beijing 100191, Peoples R China
[2] Beihang Univ, Sch Cyber Sci & Technol, Beijing 100191, Peoples R China
[3] Tsinghua Univ, Dept Comp Sci & Technol, Beijing 100084, Peoples R China
基金
中国国家自然科学基金;
关键词
Tensors; Sparse matrices; Hardware; Convolutional neural networks; Matrix decomposition; Unsupervised learning; Task analysis; Tensor computation; MTTKRP; sparse tensor storage format; convolutional neural network; convolutional autoencoder; DECOMPOSITIONS;
D O I
10.1109/TC.2021.3113028
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Canonical polyadic decomposition (CPD) is one of the most common tensor computations adopted in many scientific applications. The major bottleneck of CPD is matricized tensor times Khatri-Rao product (MTTKRP). To optimize the performance of MTTKRP, various sparse tensor formats have been proposed such as CSF and HiCOO. However, due to the spatial complexity of the tensors, no single format fits all tensors. To address this problem, we propose SpTFS, a framework that automatically predicts the optimal storage format for an input sparse tensor. Specifically, SpTFS leverages a set of sampling methods to lower the sparse tensor to fixed-sized matrices and sparsity features. In addition, SpTFS adopts both supervised learning based and unsupervised learning based methods to predict the optimal sparse tensor storage formats. For supervised learning, we propose TnsNet that combines convolution neural network (CNN) and the feature layer, which effectively captures the sparsity patterns of the input tensors. Once trained, the TnsNet can be used with either density or histogram representation of the input tensor for optimal format prediction. Whereas for unsupervised learning, we propose TnsClustering that consists of a feature encoder using convolutional layers and fully connected layers, and a K-means++ model to cluster sparse tensors for optimal tensor format prediction, without massively profiling on the hardware platform. SpTFS can use the above two models to predict the optimal tensor storage format for accelerating MTTKRP accurately. The experimental results show that both TnsNet and TnsClustering can achieve higher prediction accuracy and performance speedup compared to the state-of-the-art works.
引用
收藏
页码:1968 / 1981
页数:14
相关论文
共 8 条
  • [1] Input-Aware Sparse Tensor Storage Format Selection for Optimizing MTTKRP
    Yang, Hailong
    Liu, Yi
    Luan, Zhongzhi
    Gan, Lin
    Yang, Guangwen
    Qian, Depei
    COMPUTER, 2023, 56 (08) : 4 - 7
  • [2] SpTFS: Sparse Tensor Format Selection for MTTKRP via Deep Learning
    Sun, Qingxiao
    Liu, Yi
    Dun, Ming
    Yang, Hailong
    Luan, Zhongzhi
    Gan, Lin
    Yang, Guangwen
    Qian, Depei
    PROCEEDINGS OF SC20: THE INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS (SC20), 2020,
  • [3] An Input-aware Factorization Machine for Sparse Prediction
    Yu, Yantao
    Wang, Zhen
    Yuan, Bo
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 1466 - 1472
  • [4] Sparsity-Aware Storage Format Selection
    Cheshmi, Kazem
    Cheshmi, Leila
    Dehnavi, Maryam Mehri
    PROCEEDINGS 2018 INTERNATIONAL CONFERENCE ON HIGH PERFORMANCE COMPUTING & SIMULATION (HPCS), 2018, : 1034 - 1037
  • [5] Input-Aware Implication Selection Scheme Utilizing ATPG for Efficient Concurrent Error Detection
    Hassan, Abdus Sami
    Afzaal, Umar
    Arifeen, Tooba
    Lee, Jeong A.
    ELECTRONICS, 2018, 7 (10)
  • [6] Optimizing Sparse Linear Algebra Through Automatic Format Selection and Machine Learning
    Stylianou, Christodoulos
    Weiland, Michele
    2023 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS, IPDPSW, 2023, : 734 - 743
  • [7] IA-SpGEMM An Input-aware Auto-tuning Framework for Parallel Sparse Matrix-Matrix Multiplication
    Xie, Zhen
    Tan, Guangming
    Liu, Weifeng
    Sun, Ninghui
    INTERNATIONAL CONFERENCE ON SUPERCOMPUTING (ICS 2019), 2019, : 94 - 105
  • [8] WACO: LearningWorkload-Aware Co-optimization of the Format and Schedule of a Sparse Tensor Program
    Won, Jaeyeon
    Mendis, Charith
    Emer, Joel S.
    Amarasinghe, Saman
    PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON ARCHITECTURAL SUPPORT FOR PROGRAMMING LANGUAGES AND OPERATING SYSTEMS, VOL 2, ASPLOS 2023, 2023, : 920 - 934