Software Defects Identification: Results Using Machine Learning and Explainable Artificial Intelligence Techniques

被引:3
|
作者
Begum, Momotaz [1 ]
Shuvo, Mehedi Hasan [1 ]
Ashraf, Imran [2 ]
Al Mamun, Abdullah
Uddin, Jia [3 ]
Samad, Md Abdus [2 ]
机构
[1] Dhaka Univ Engn & Technol DUET, Dept Comp Sci & Engn, Gazipur 1707, Bangladesh
[2] Yeungnam Univ, Dept Informat & Commun Engn, Gyongsan 38541, Gyeongsangbuk D, South Korea
[3] Woosong Univ, Endicott Coll, AI & Big Data Dept, Daejeon 34606, South Korea
关键词
Software reliability; Fault diagnosis; Artificial intelligence; Unified modeling language; Random forests; Predictive models; Software defect prediction; features selection; software reliability; software fault diagnosis; explainable AI; SHAP; LIME; TIME;
D O I
10.1109/ACCESS.2023.3329051
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The rising deployment of software in automation and the cognitive skills of machines indicate a machine revolution in modern human civilization. Thus, diagnosing and predicting software faults is crucial to software reliability. In this paper, we first preprocessed four real datasets offered by National Aeronautics and Space Administration with twenty-one features using the Synthetic Minority Oversampling Technique and Label Encoding techniques. Subsequently, we experimented with thirteen software fault diagnosis Machine Learning (ML) models, i.e., Random Forest Regression, Linear Regression, Naive Bayes, Decision Tree Classifier, Logistic Regression, KNeighbors Classifier, AdaBoost, Gradient Boosting Classifier, Gradient Boosting Regression, XGBR Regressor, XGBoost Classifier, Extra Trees Classifier and Support Vectors Machine after that, we compared each ML Model to select the best diagnostic model. Among them, XGBR outperformed, considering the accuracy, mean square error, and R2 score. We also used Explainable Artificial Intelligence (XAI), Local Interpretable Model (LIME), and SHapley Additive exPlanations (SHAP) to determine software fault features. We observed that Number of static invocations (nosi), Depth Inheritance Tree (dit), and Coupling Between Objects (cbo) features are the most affected software faults feature from datasets. For LIME, the average True positive of nosi is 40%, dit is 15%, and cbo is 20%; on the other hand, the SHAP average true positive value of nosi is 36%, cbo is 15%, and the norm true negative value of dit is 5%. Thus, LIME can afford the greatest impact on the model outcomes to identify features that are the most significant reasons for software defects.
引用
收藏
页码:132750 / 132765
页数:16
相关论文
共 50 条
  • [41] Machine learning-based prediction of Clostridium growth in pork meat using explainable artificial intelligence
    Ince, Volkan
    Bader-El-Den, Mohamed
    Alderton, Jack
    Arabikhan, Farzad
    Sari, Omer Faruk
    Sansom, Annette
    JOURNAL OF FOOD SCIENCE AND TECHNOLOGY-MYSORE, 2025,
  • [42] Modeling and Predictive Analytics of Breast Cancer Using Ensemble Learning Techniques: An Explainable Artificial Intelligence Approach
    Raha, Avi Deb
    Dihan, Fatema Jannat
    Gain, Mrityunjoy
    Murad, Saydul Akbar
    Adhikary, Apurba
    Hossain, Md. Bipul
    Hassan, Md. Mehedi
    Al-Shehari, Taher
    Alsadhan, Nasser A.
    Kadrie, Mohammed
    Bairagi, Anupam Kumar
    Computers, Materials and Continua, 2024, 81 (03): : 4033 - 4048
  • [43] An Explainable Artificial Intelligence Framework for the Predictive Analysis of Hypo and Hyper Thyroidism Using Machine Learning Algorithms
    Md. Bipul Hossain
    Anika Shama
    Apurba Adhikary
    Avi Deb Raha
    K. M. Aslam Uddin
    Mohammad Amzad Hossain
    Imtia Islam
    Saydul Akbar Murad
    Md. Shirajum Munir
    Anupam Kumar Bairagi
    Human-Centric Intelligent Systems, 2023, 3 (3): : 211 - 231
  • [44] An explainable Artificial Intelligence software system for predicting diabetes
    Srinivasu, Parvathaneni Naga
    Ahmed, Shakeel
    Hassaballah, Mahmoud
    Almusallam, Naif
    HELIYON, 2024, 10 (16)
  • [45] Identification of biological markers in cancer disease using explainable artificial intelligence
    Shahzad, Muhammad
    Lohana, Ruhal
    Aurangzeb, Khursheed
    Ali, Isbah Imtiaz
    Anwar, Muhammad Shahid
    Murtaza, Mahnoor
    Malick, Rauf Ahmed Shams
    Allayarov, Piratdin
    INTERNATIONAL JOURNAL OF IMAGING SYSTEMS AND TECHNOLOGY, 2024, 34 (02)
  • [46] Explainable artificial intelligence for stroke prediction through comparison of deep learning and machine learning models
    Moulaei, Khadijeh
    Afshari, Lida
    Moulaei, Reza
    Sabet, Babak
    Mousavi, Seyed Mohammad
    Afrash, Mohammad Reza
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [47] Gender identification of chicks using vocalisation signals, artificial intelligence and machine learning techniques: current status and future prospects
    Kannan, A.
    Basu, Joyanta
    Roy, Rajib
    Pal, Madhab
    Rao, S. V. Rama
    Chatterjee, R. N.
    Ghosh, Tarun
    Ray, Hena
    Ghosh, Alokesh
    WORLDS POULTRY SCIENCE JOURNAL, 2025, 81 (01) : 87 - 102
  • [48] The Use of eXplainable Artificial Intelligence and Machine Learning Operation Principles to Support the Continuous Development of Machine Learning-Based Solutions in Fault Detection and Identification
    Tran, Tuan-Anh
    Ruppert, Tamas
    Abonyi, Janos
    COMPUTERS, 2024, 13 (10)
  • [49] Diabetes prediction using machine learning and explainable AI techniques
    Tasin, Isfafuzzaman
    Nabil, Tansin Ullah
    Islam, Sanjida
    Khan, Riasat
    HEALTHCARE TECHNOLOGY LETTERS, 2023, 10 (1-2) : 1 - 10
  • [50] Interactive Collaborative Learning with Explainable Artificial Intelligence
    Arnold, Oksana
    Golchert, Sebastian
    Rennert, Michel
    Jantke, Klaus P.
    LEARNING IN THE AGE OF DIGITAL AND GREEN TRANSITION, ICL2022, VOL 1, 2023, 633 : 13 - 24