Software Defects Identification: Results Using Machine Learning and Explainable Artificial Intelligence Techniques

被引:3
|
作者
Begum, Momotaz [1 ]
Shuvo, Mehedi Hasan [1 ]
Ashraf, Imran [2 ]
Al Mamun, Abdullah
Uddin, Jia [3 ]
Samad, Md Abdus [2 ]
机构
[1] Dhaka Univ Engn & Technol DUET, Dept Comp Sci & Engn, Gazipur 1707, Bangladesh
[2] Yeungnam Univ, Dept Informat & Commun Engn, Gyongsan 38541, Gyeongsangbuk D, South Korea
[3] Woosong Univ, Endicott Coll, AI & Big Data Dept, Daejeon 34606, South Korea
关键词
Software reliability; Fault diagnosis; Artificial intelligence; Unified modeling language; Random forests; Predictive models; Software defect prediction; features selection; software reliability; software fault diagnosis; explainable AI; SHAP; LIME; TIME;
D O I
10.1109/ACCESS.2023.3329051
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The rising deployment of software in automation and the cognitive skills of machines indicate a machine revolution in modern human civilization. Thus, diagnosing and predicting software faults is crucial to software reliability. In this paper, we first preprocessed four real datasets offered by National Aeronautics and Space Administration with twenty-one features using the Synthetic Minority Oversampling Technique and Label Encoding techniques. Subsequently, we experimented with thirteen software fault diagnosis Machine Learning (ML) models, i.e., Random Forest Regression, Linear Regression, Naive Bayes, Decision Tree Classifier, Logistic Regression, KNeighbors Classifier, AdaBoost, Gradient Boosting Classifier, Gradient Boosting Regression, XGBR Regressor, XGBoost Classifier, Extra Trees Classifier and Support Vectors Machine after that, we compared each ML Model to select the best diagnostic model. Among them, XGBR outperformed, considering the accuracy, mean square error, and R2 score. We also used Explainable Artificial Intelligence (XAI), Local Interpretable Model (LIME), and SHapley Additive exPlanations (SHAP) to determine software fault features. We observed that Number of static invocations (nosi), Depth Inheritance Tree (dit), and Coupling Between Objects (cbo) features are the most affected software faults feature from datasets. For LIME, the average True positive of nosi is 40%, dit is 15%, and cbo is 20%; on the other hand, the SHAP average true positive value of nosi is 36%, cbo is 15%, and the norm true negative value of dit is 5%. Thus, LIME can afford the greatest impact on the model outcomes to identify features that are the most significant reasons for software defects.
引用
收藏
页码:132750 / 132765
页数:16
相关论文
共 50 条
  • [21] Explainable artificial intelligence and interpretable machine learning for agricultural data analysis
    Ryo, Masahiro
    ARTIFICIAL INTELLIGENCE IN AGRICULTURE, 2022, 6 : 257 - 265
  • [22] Machine learning and explainable artificial intelligence for the prevention of waterborne cryptosporidiosis and giardiosis
    Ligda, Panagiota
    Mittas, Nikolaos
    Kyzas, George Z.
    Claerebout, Edwin
    Sotiraki, Smaragda
    WATER RESEARCH, 2024, 262
  • [23] Explainable artificial intelligence and machine learning algorithms for classification of thyroid disease
    Kumari, Priyanka
    Kaur, Baljinder
    Rakhra, Manik
    Deka, Aniruddha
    Byeon, Haewon
    Asenso, Evans
    Rawat, Anil Kumar
    DISCOVER APPLIED SCIENCES, 2024, 6 (07)
  • [24] A Machine Learning and Explainable Artificial Intelligence Approach for Insurance Fraud Classification
    Khan, Zaid
    Olivia, Diana
    Shetty, Sucharita
    INTELIGENCIA ARTIFICIAL-IBEROAMERICAN JOURNAL OF ARTIFICIAL INTELLIGENCE, 2025, 28 (75): : 140 - 169
  • [25] A decision support system for osteoporosis risk prediction using machine learning and explainable artificial intelligence
    Khanna, Varada Vivek
    Chadaga, Krishnaraj
    Sampathila, Niranjana
    Chadaga, Rajagopala
    Prabhu, Srikanth
    Swathi, K. S.
    Jagdale, Aditya S.
    Bhat, Devadas
    HELIYON, 2023, 9 (12)
  • [26] Unraveling the Impact of Land Cover Changes on Climate Using Machine Learning and Explainable Artificial Intelligence
    Kolevatova, Anastasiia
    Riegler, Michael A.
    Cherubini, Francesco
    Hu, Xiangping
    Hammer, Hugo L.
    BIG DATA AND COGNITIVE COMPUTING, 2021, 5 (04)
  • [27] Interpretable Machine Learning Models for Malicious Domains Detection Using Explainable Artificial Intelligence (XAI)
    Aslam, Nida
    Khan, Irfan Ullah
    Mirza, Samiha
    AlOwayed, Alanoud
    Anis, Fatima M.
    Aljuaid, Reef M.
    Baageel, Reham
    SUSTAINABILITY, 2022, 14 (12)
  • [28] Basic Artificial Intelligence Techniques Machine Learning and Deep Learning
    Erickson, Bradley J.
    RADIOLOGIC CLINICS OF NORTH AMERICA, 2021, 59 (06) : 933 - 940
  • [29] On Using Explainable Artificial Intelligence for Failure Identification in Microwave Networks
    Ayoub, Omran
    Musumeci, Francesco
    Ezzeddine, Fatima
    Passera, Claudio
    Tornatore, Massimo
    25TH CONFERENCE ON INNOVATION IN CLOUDS, INTERNET AND NETWORKS (ICIN 2022), 2022, : 48 - 55
  • [30] Enhancing Intrusion Detection Systems With Advanced Machine Learning Techniques: An Ensemble and Explainable Artificial Intelligence (AI) Approach
    Alatawi, Mohammed Naif
    SECURITY AND PRIVACY, 2025, 8 (01):