ReACT: Redundancy-Aware Code Generation for Tensor Expressions

被引:0
|
作者
Zhou, Tong [1 ]
Tian, Ruiqin [2 ]
Ashraf, Rizwan A. [3 ]
Gioiosa, Roberto [3 ]
Kestor, Gokcen [3 ]
Sarkar, Vivek [1 ]
机构
[1] Georgia Inst Technol, Atlanta, GA 30332 USA
[2] Horizon Robot, Shanghai, Peoples R China
[3] Pacif Northwest Natl Lab, Richland, WA USA
关键词
D O I
10.1145/3559009.3569685
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
High-level programming models for tensor computations are becoming increasingly popular in many domains such as machine learning and data science. The index notation is one such model that is widely adopted for expressing a wide range of tensor computations algorithmically and also as input to programming systems. In programming systems, sparse tensors can be specified as type annotations, and a compiler can be employed to perform code generation for the specified tensor expressions and sparse formats. Different code generation strategies and optimization decisions can have a significant impact on the performance of the generated code. However, the code generation strategies used by current state-of-the-art tensor compilers can result in redundant computations being present in the output code. In this work, we identify four common types of redundancies that can occur when generating code for compound expressions, and introduce new techniques that can avoid these redundancies. Empirical evaluation on real-world compound kernels, such as Sampled Dense Dense Matrix Multiplication (SDDMM), Graph Neural Network (GNN) and Matricized-Tensor Times Khatri-Rao Product (MTTKRP) shows that our generated code with redundancy elimination can result in performance improvements of 1.1x to 25x relative to a state-of-the-art Tensor Algebra COmpiler (TACO) and up to 101x relative to library approaches such as the SciPy.sparse.
引用
收藏
页码:1 / 13
页数:13
相关论文
共 50 条
  • [1] Redundancy-Aware Maximal Cliques
    Wang, Jia
    Cheng, James
    Fu, Ada Wai-Chee
    [J]. 19TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING (KDD'13), 2013, : 122 - 130
  • [2] Redundancy-Aware Action Spaces for Robot Learning
    Mazzaglia, Pietro
    Backshall, Nicholas
    Ma, Xiao
    James, Stephen
    [J]. IEEE ROBOTICS AND AUTOMATION LETTERS, 2024, 9 (08): : 6912 - 6919
  • [3] Redundancy-aware Transformer for Video Question Answering
    Li, Yicong
    Yang, Xun
    Zhang, An
    Feng, Chun
    Wang, Xiang
    Chua, Tat-Seng
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 3172 - 3180
  • [4] Redundancy-Aware Pruning of Convolutional Neural Networks
    Xie, Guotian
    [J]. NEURAL COMPUTATION, 2020, 32 (12) : 2482 - 2506
  • [5] Redundancy-aware topology management in wireless sensor networks
    Al-Omari, Safwan
    Shi, Weisong
    [J]. 2006 INTERNATIONAL CONFERENCE ON COLLABORATIVE COMPUTING: NETWORKING, APPLICATIONS AND WORKSHARING, 2006, : 29 - +
  • [6] A Redundancy-Aware Face Structure for Wireless Sensor Networks
    Razzaq, Ammara
    Khedr, Ahmed M.
    Al Aghbari, Zaher
    [J]. 2018 8TH INTERNATIONAL CONFERENCE ON COMPUTER SCIENCE AND INFORMATION TECHNOLOGY (CSIT), 2018, : 38 - 42
  • [7] Redundancy-Aware Electromigration Checking for Mesh Power Grids
    Chatterjee, Sandeep
    Fawaz, Mohammad
    Najm, Farid N.
    [J]. 2013 IEEE/ACM INTERNATIONAL CONFERENCE ON COMPUTER-AIDED DESIGN (ICCAD), 2013, : 540 - 547
  • [8] Redundancy-Aware Topic Modeling for Patient Record Notes
    Cohen, Raphael
    Aviram, Iddo
    Elhadad, Michael
    Elhadad, Noemie
    [J]. PLOS ONE, 2014, 9 (02):
  • [9] Fast CNN Pruning via Redundancy-Aware Training
    Dong, Xiao
    Liu, Lei
    Li, Guangli
    Zhao, Peng
    Feng, Xiaobing
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2018, PT I, 2018, 11139 : 3 - 13
  • [10] Towards a Redundancy-Aware Network Stack for Data Centers
    Iftikhar, Ali Musa
    Dogar, Fahad R.
    Qazi, Ihsan Ayyub
    [J]. PROCEEDINGS OF THE 15TH ACM WORKSHOP ON HOT TOPICS IN NETWORKS (HOTNETS '16), 2016, : 57 - 63