An Empirical Study on Common Bugs in Deep Learning Compilers

被引:3
|
作者
Du, Xiaoting [1 ]
Zheng, Zheng [1 ]
Ma, Lei [2 ]
Zhao, Jianjun [3 ]
机构
[1] Beihang Univ, Beijing, Peoples R China
[2] Univ Alberta, Edmonton, AB, Canada
[3] Kyushu Univ, Fukuoka, Japan
基金
中国国家自然科学基金;
关键词
empirical study; deep learning compiler; bug; root cause; impact;
D O I
10.1109/ISSRE52982.2021.00030
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
The highly diversified deep learning (DL) frameworks and target hardware architectures bring big challenges for DL model deployment for industrial production. Up to the present, continuous efforts have been made to develop DL compilers with multiple state-of-the-arts available, e.g., TVM, Glow, nGraph, PlaidML, and Tensor Comprehensions (TC). Unlike traditional compilers, DL compilers take a DL model built by DL frameworks as input and generate optimized code as the output for a particular target device. Similar to other software, DL compilers are also error-prone. Buggy DL compilers can generate incorrect code and result in unexpected model behaviors. To better understand the current status and common bug characteristics of DL compilers, we performed a large-scale empirical study of five popular DL compilers covering TVM, Glow, nGraph, PlaidML, and TC, collecting a total of 2,717 actual bug reports submitted by users and developers. We made large manual efforts to investigate these bug reports and classified them based on their root causes, during which five root causes were identified, including environment, compatibility, memory, document, and semantic. After labeling the types of bugs, we further examined the important consequences of each type of bug and analyzed the correlation between bug types and impacts. Besides, we studied the time required to fix different types of bugs in DL compilers. Seven important findings are eventually obtained, with practical implications provided for both DL compiler developers and users.
引用
收藏
页码:184 / 195
页数:12
相关论文
共 50 条
  • [21] Systematic Study of Compilers and Vulnerability Scanners Using the Example of Integer Bugs
    di Angelo, Monika
    Mohr, Rafael
    Salzer, Gernot
    2024 IEEE INTERNATIONAL CONFERENCE ON BLOCKCHAIN, BLOCKCHAIN 2024, 2024, : 243 - 251
  • [22] Active learning: an empirical study of common baselines
    Ramirez-Loaiza, Maria E.
    Sharma, Manali
    Kumar, Geet
    Bilgic, Mustafa
    DATA MINING AND KNOWLEDGE DISCOVERY, 2017, 31 (02) : 287 - 313
  • [23] Active learning: an empirical study of common baselines
    Maria E. Ramirez-Loaiza
    Manali Sharma
    Geet Kumar
    Mustafa Bilgic
    Data Mining and Knowledge Discovery, 2017, 31 : 287 - 313
  • [24] Understanding and Detecting Performance Bugs in Markdown Compilers
    Li, Penghui
    Liu, Yinxi
    Meng, Wei
    2021 36TH IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING ASE 2021, 2021, : 892 - 904
  • [25] One-Shot Tuner for Deep Learning Compilers
    Ryu, Jaehun
    Park, Eunhyeok
    Sung, Hyojin
    CC'22: PROCEEDINGS OF THE 31ST ACM SIGPLAN INTERNATIONAL CONFERENCE ON COMPILER CONSTRUCTION, 2022, : 89 - 103
  • [26] Performance Evaluation of Deep Learning Compilers for Edge Inference
    Verma, Gaurav
    Gupta, Yashi
    Malik, Abid M.
    Chapman, Barbara
    2021 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS (IPDPSW), 2021, : 858 - 865
  • [27] ALT : Optimizing Tensor Compilation in Deep Learning Compilers with Active Learning
    Zeng, Xi
    Zhi, Tian
    Du, Zidong
    Guo, Qi
    Sun, Ninghui
    Chen, Yunji
    2020 IEEE 38TH INTERNATIONAL CONFERENCE ON COMPUTER DESIGN (ICCD 2020), 2020, : 623 - 630
  • [28] An Empirical Study on Critical Blocking Bugs
    Ren, Hao
    Li, Yanhui
    Chen, Lin
    2020 IEEE/ACM 28TH INTERNATIONAL CONFERENCE ON PROGRAM COMPREHENSION, ICPC, 2020, : 72 - 82
  • [29] An Empirical Study on TensorFlow Program Bugs
    Zhang, Yuhao
    Chen, Yifan
    Cheung, Shing-Chi
    Xiong, Yingfei
    Zhang, Lu
    ISSTA'18: PROCEEDINGS OF THE 27TH ACM SIGSOFT INTERNATIONAL SYMPOSIUM ON SOFTWARE TESTING AND ANALYSIS, 2018, : 129 - 140
  • [30] An Empirical Study of Bugs in Test Code
    Vahabzadeh, Arash
    Fard, Amin Milani
    Mesbah, Ali
    2015 31ST INTERNATIONAL CONFERENCE ON SOFTWARE MAINTENANCE AND EVOLUTION (ICSME) PROCEEDINGS, 2015, : 101 - 110