Explaining anomalies detected by autoencoders using Shapley Additive Explanations

被引:117
|
作者
Antwarg, Liat [1 ]
Miller, Ronnie Mindlin [1 ]
Shapira, Bracha [1 ]
Rokach, Lior [1 ]
机构
[1] Ben Gurion Univ Negev, Dept Informat & Software Syst Engn, Beer Sheva, Israel
关键词
Explainable black-box models; XAI; Autoencoder; Shapley values; SHAP; Anomaly detection; NETWORK;
D O I
10.1016/j.eswa.2021.115736
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning algorithms for anomaly detection, such as autoencoders, point out the outliers, saving experts the time-consuming task of examining normal cases in order to find anomalies. Most outlier detection algorithms output a score for each instance in the database. The top-k most intense outliers are returned to the user for further inspection; however, the manual validation of results becomes challenging without justification or additional clues. An explanation of why an instance is anomalous enables the experts to focus their investigation on the most important anomalies and may increase their trust in the algorithm. Recently, a game theory-based framework known as SHapley Additive exPlanations (SHAP) was shown to be effective in explaining various supervised learning models. In this paper, we propose a method that uses Kernel SHAP to explain anomalies detected by an autoencoder, which is an unsupervised model. The proposed explanation method aims to provide a comprehensive explanation to the experts by focusing on the connection between the features with high reconstruction error and the features that are most important in terms of their affect on the reconstruction error. We propose a black-box explanation method, because it has the advantage of being able to explain any autoencoder without being aware of the exact architecture of the autoencoder model. The proposed explanation method extracts and visually depicts both features that contribute the most to the anomaly and those that offset it. An expert evaluation using real-world data demonstrates the usefulness of the proposed method in helping domain experts better understand the anomalies. Our evaluation of the explanation method, in which a "perfect"autoencoder is used as the ground truth, shows that the proposed method explains anomalies correctly, using the exact features, and evaluation on real-data demonstrates that (1) our explanation model, which uses SHAP, is more robust than the Local Interpretable Model-agnostic Explanations (LIME) method, and (2) the explanations our method provides are more effective at reducing the anomaly score than other methods.
引用
下载
收藏
页数:14
相关论文
共 50 条
  • [1] RESHAPE: Explaining Accounting Anomalies in Financial Statement Audits by enhancing SHapley Additive exPlanations
    Muller, Ricardo
    Schreyer, Marco
    Sattarov, Timur
    Borth, Damian
    3RD ACM INTERNATIONAL CONFERENCE ON AI IN FINANCE, ICAIF 2022, 2022, : 174 - 182
  • [2] Index tracking using shapley additive explanations and one-dimensional pointwise convolutional autoencoders
    Zhang, Yanyi
    De Smedt, Johannes
    INTERNATIONAL REVIEW OF FINANCIAL ANALYSIS, 2024, 95
  • [3] Explaining deep learning-based activity schedule models using SHapley Additive exPlanations
    Koushik, Anil
    Manoj, M.
    Nezamuddin, N.
    TRANSPORTATION LETTERS-THE INTERNATIONAL JOURNAL OF TRANSPORTATION RESEARCH, 2024,
  • [4] EXPLAINING DEEP LEARNING MODELS FOR SPOOFING AND DEEPFAKE DETECTION WITH SHAPLEY ADDITIVE EXPLANATIONS
    Ge, Wanying
    Patino, Jose
    Todisco, Massimiliano
    Evans, Nicholas
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 6387 - 6391
  • [5] Explaining Deep Q-Learning Experience Replay with SHapley Additive exPlanations
    Sullivan, Robert S.
    Longo, Luca
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2023, 5 (04): : 1433 - 1455
  • [6] Explaining Intrusion Detection-Based Convolutional Neural Networks Using Shapley Additive Explanations (SHAP)
    Younisse, Remah
    Ahmad, Ashraf
    Abu Al-Haija, Qasem
    BIG DATA AND COGNITIVE COMPUTING, 2022, 6 (04)
  • [7] Additive Explanations for Anomalies Detected from Multivariate Temporal Data
    Giurgiu, Ioana
    Schumann, Anika
    PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT (CIKM '19), 2019, : 2245 - 2248
  • [8] SHapley Additive exPlanations for Explaining Artificial Neural Network Based Mode Choice Models
    Anil Koushik
    M. Manoj
    N. Nezamuddin
    Transportation in Developing Economies, 2024, 10
  • [9] SHapley Additive exPlanations for Explaining Artificial Neural Network Based Mode Choice Models
    Koushik, Anil
    Manoj, M.
    Nezamuddin, N.
    TRANSPORTATION IN DEVELOPING ECONOMIES, 2024, 10 (01)
  • [10] Bankruptcy prediction using machine learning and Shapley additive explanations
    Nguyen, Hoang Hiep
    Viviani, Jean-Laurent
    Ben Jabeur, Sami
    REVIEW OF QUANTITATIVE FINANCE AND ACCOUNTING, 2023,