Interpretable machine learning with tree-based shapley additive explanations: Application to metabolomics datasets for binary classification

被引:25
|
作者
Bifarin, Olatomiwa O. [1 ,2 ]
机构
[1] Univ Georgia, Dept Biochem & Mol Biol, Athens, GA 30602 USA
[2] Georgia Inst Technol, Sch Chem & Biochem, Atlanta, GA 30602 USA
来源
PLOS ONE | 2023年 / 18卷 / 05期
关键词
METABOLIGHTS; REPOSITORY;
D O I
10.1371/journal.pone.0284315
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Machine learning (ML) models are used in clinical metabolomics studies most notably for biomarker discoveries, to identify metabolites that discriminate between a case and control group. To improve understanding of the underlying biomedical problem and to bolster confidence in these discoveries, model interpretability is germane. In metabolomics, partial least square discriminant analysis (PLS-DA) and its variants are widely used, partly due to the model's interpretability with the Variable Influence in Projection (VIP) scores, a global interpretable method. Herein, Tree-based Shapley Additive explanations (SHAP), an interpretable ML method grounded in game theory, was used to explain ML models with local explanation properties. In this study, ML experiments (binary classification) were conducted for three published metabolomics datasets using PLS-DA, random forests, gradient boosting, and extreme gradient boosting (XGBoost). Using one of the datasets, PLS-DA model was explained using VIP scores, while one of the best-performing models, a random forest model, was interpreted using Tree SHAP. The results show that SHAP has a more explanation depth than PLS-DA's VIP, making it a powerful method for rationalizing machine learning predictions from metabolomics studies.
引用
收藏
页数:21
相关论文
共 50 条
  • [1] An interpretable framework for modeling global solar radiation using tree-based ensemble machine learning and Shapley additive explanations methods
    Song, Zhe
    Cao, Sunliang
    Yang, Hongxing
    APPLIED ENERGY, 2024, 364
  • [2] Interpretable Machine Learning in Damage Detection Using Shapley Additive Explanations
    Movsessian, Artur
    Cava, David Garcia
    Tcherniak, Dmitri
    ASCE-ASME JOURNAL OF RISK AND UNCERTAINTY IN ENGINEERING SYSTEMS PART B-MECHANICAL ENGINEERING, 2022, 8 (02):
  • [3] Tree-based interpretable machine learning of the thermodynamic phases
    Yang, Jintao
    Cao, Junpeng
    PHYSICS LETTERS A, 2021, 412
  • [4] Interpretable prediction of thermal sensation for elderly people based on data sampling, machine learning and SHapley Additive exPlanations (SHAP)
    Zheng, Guozhong
    Zhang, Yuqin
    Yue, Xuhui
    Li, Kang
    BUILDING AND ENVIRONMENT, 2023, 242
  • [5] Interpretable Machine Learning Models Based on Shapley Additive Explanations for Predicting the Risk of Cerebrospinal Fluid Leakage in Lumbar Fusion Surgery
    Guo, Zongjie
    Wang, Peiyang
    Ye, Suhui
    Li, Haoyu
    Bao, Junping
    Shi, Rui
    Yang, Shu
    Yin, Rui
    Wu, Xiaotao
    SPINE, 2024, 49 (18) : 1281 - 1293
  • [6] Predicting the pathological invasiveness in patients with a solitary pulmonary nodule via Shapley additive explanations interpretation of a tree-based machine learning radiomics model: a multicenter study
    Zhang, Rong
    Hong, Minping
    Cai, Hongjie
    Liang, Yanting
    Chen, Xinjie
    Liu, Ziwei
    Wu, Meilian
    Zhou, Cuiru
    Bao, Chenzhengren
    Wang, Huafeng
    Yang, Shaomin
    Hu, Qiugen
    QUANTITATIVE IMAGING IN MEDICINE AND SURGERY, 2023, 13 (12) : 7828 - +
  • [7] Bankruptcy prediction using machine learning and Shapley additive explanations
    Nguyen, Hoang Hiep
    Viviani, Jean-Laurent
    Ben Jabeur, Sami
    REVIEW OF QUANTITATIVE FINANCE AND ACCOUNTING, 2023,
  • [8] Interpretable prediction of cardiopulmonary complications after non-small cell lung cancer surgery based on machine learning and SHapley additive exPlanations
    Zhai, Yihai
    Lin, Xue
    Wei, Qiaolin
    Pu, Yuanjin
    Pang, Yonghui
    HELIYON, 2023, 9 (07)
  • [9] Developing interpretable machine learning-Shapley additive explanations model for unconfined compressive strength of cohesive soils stabilized with geopolymer
    Ngo, Anh Quan
    Nguyen, Linh Quy
    Tran, Van Quan
    PLOS ONE, 2023, 18 (06):
  • [10] A Novel Tree-Based Method for Interpretable Reinforcement Learning
    Li, Yifan
    Qi, Shuhan
    Wang, Xuan
    Zhang, Jiajia
    Cui, Lei
    ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2024, 18 (09)