Software Defects Identification: Results Using Machine Learning and Explainable Artificial Intelligence Techniques

被引:3
|
作者
Begum, Momotaz [1 ]
Shuvo, Mehedi Hasan [1 ]
Ashraf, Imran [2 ]
Al Mamun, Abdullah
Uddin, Jia [3 ]
Samad, Md Abdus [2 ]
机构
[1] Dhaka Univ Engn & Technol DUET, Dept Comp Sci & Engn, Gazipur 1707, Bangladesh
[2] Yeungnam Univ, Dept Informat & Commun Engn, Gyongsan 38541, Gyeongsangbuk D, South Korea
[3] Woosong Univ, Endicott Coll, AI & Big Data Dept, Daejeon 34606, South Korea
关键词
Software reliability; Fault diagnosis; Artificial intelligence; Unified modeling language; Random forests; Predictive models; Software defect prediction; features selection; software reliability; software fault diagnosis; explainable AI; SHAP; LIME; TIME;
D O I
10.1109/ACCESS.2023.3329051
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The rising deployment of software in automation and the cognitive skills of machines indicate a machine revolution in modern human civilization. Thus, diagnosing and predicting software faults is crucial to software reliability. In this paper, we first preprocessed four real datasets offered by National Aeronautics and Space Administration with twenty-one features using the Synthetic Minority Oversampling Technique and Label Encoding techniques. Subsequently, we experimented with thirteen software fault diagnosis Machine Learning (ML) models, i.e., Random Forest Regression, Linear Regression, Naive Bayes, Decision Tree Classifier, Logistic Regression, KNeighbors Classifier, AdaBoost, Gradient Boosting Classifier, Gradient Boosting Regression, XGBR Regressor, XGBoost Classifier, Extra Trees Classifier and Support Vectors Machine after that, we compared each ML Model to select the best diagnostic model. Among them, XGBR outperformed, considering the accuracy, mean square error, and R2 score. We also used Explainable Artificial Intelligence (XAI), Local Interpretable Model (LIME), and SHapley Additive exPlanations (SHAP) to determine software fault features. We observed that Number of static invocations (nosi), Depth Inheritance Tree (dit), and Coupling Between Objects (cbo) features are the most affected software faults feature from datasets. For LIME, the average True positive of nosi is 40%, dit is 15%, and cbo is 20%; on the other hand, the SHAP average true positive value of nosi is 36%, cbo is 15%, and the norm true negative value of dit is 5%. Thus, LIME can afford the greatest impact on the model outcomes to identify features that are the most significant reasons for software defects.
引用
收藏
页码:132750 / 132765
页数:16
相关论文
共 50 条
  • [1] Explainable Artificial Intelligence and Machine Learning
    Raunak, M. S.
    Kuhn, Rick
    COMPUTER, 2021, 54 (10) : 25 - 27
  • [2] Explainable Software Defects Classification Using SMOTE and Machine Learning
    Jude A.
    Uddin J.
    Annals of Emerging Technologies in Computing, 2024, 8 (01) : 35 - 49
  • [3] Prediction of disease comorbidity using explainable artificial intelligence and machine learning techniques: A systematic review
    Alsaleh, Mohanad M.
    Allery, Freya
    Choi, Jung Won
    Hama, Tuankasfee
    McQuillin, Andrew
    Wu, Honghan
    Thygesen, Johan H.
    INTERNATIONAL JOURNAL OF MEDICAL INFORMATICS, 2023, 175
  • [4] Predicting Software Defects with Explainable Machine Learning
    Santos, Geanderson
    Figueiredo, Eduardo
    Veloso, Adriano
    Viggiato, Markos
    Ziviani, Nivio
    PROCEEDINGS OF THE 19TH BRAZILIAN SYMPOSIUM ON SOFTWARE QUALITY, SBOS 2020, 2020,
  • [5] Detection of Manufacturing Defects in Steel Using Deep Learning With Explainable Artificial Intelligence
    Aboulhosn, Zeina
    Musamih, Ahmad
    Salah, Khaled
    Jayaraman, Raja
    Omar, Mohammed
    Aung, Zeyar
    IEEE ACCESS, 2024, 12 : 99240 - 99257
  • [6] Predicting compressive strength of hollow concrete prisms using machine learning techniques and explainable artificial intelligence (XAI)
    Bin Inqiad, Waleed
    Dumitrascu, Elena Valentina
    Dobre, Robert Alexandru
    Khan, Naseer Muhammad
    Hammood, Abbas Hussein
    Henedy, Sadiq N.
    Khan, Rana Muhammad Asad
    HELIYON, 2024, 10 (17)
  • [7] A Comprehensive framework for Parkinson's disease diagnosis using explainable artificial intelligence empowered machine learning techniques
    Priyadharshini, S.
    Ramkumar, K.
    Vairavasundaram, Subramaniyaswamy
    Narasimhan, K.
    Venkatesh, S.
    Amirtharajan, Rengarajan
    Kotecha, Ketan
    ALEXANDRIA ENGINEERING JOURNAL, 2024, 107 : 568 - 582
  • [8] An interpretable schizophrenia diagnosis framework using machine learning and explainable artificial intelligence
    Shivaprasad, Samhita
    Chadaga, Krishnaraj
    Dias, Cifha Crecil
    Sampathila, Niranjana
    Prabhu, Srikanth
    SYSTEMS SCIENCE & CONTROL ENGINEERING, 2024, 12 (01)
  • [9] Software defect identification using machine learning techniques
    Ceylan, Evren
    Kudubay, F. Onur
    Bener, Ayse B.
    32ND EUROMICRO CONFERENCE ON SOFTWARE ENGINEERING AND ADVANCED APPLICATIONS (SEAA) - PROCEEDINGS, 2006, : 240 - +
  • [10] Explainable artificial intelligence for machine learning prediction of bandgap energies
    Masuda, Taichi
    Tanabe, Katsuaki
    JOURNAL OF APPLIED PHYSICS, 2024, 136 (17)