Linear iterative feature embedding: an ensemble framework for an interpretable model

被引:0
|
作者
Sudjianto, Agus [1 ]
Qiu, Jinwen [1 ]
Li, Miaoqi [2 ]
Chen, Jie [3 ]
机构
[1] Wells Fargo, 401 S Tryon St, Charlotte, NC 28202 USA
[2] Wells Fargo, 11625 N Community House Rd, Charlotte, NC 28277 USA
[3] Well Fargo, 3440 Walnut Ave, Fremont, CA 94538 USA
来源
NEURAL COMPUTING & APPLICATIONS | 2023年 / 35卷 / 13期
关键词
Linear iterative feature embedding; Ensemble method; Loss decomposition; Variable importance; Interaction detection; DIVERSITY; LINK;
D O I
10.1007/s00521-023-08204-w
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
A new ensemble framework for an interpretable model called linear iterative feature embedding (LIFE) has been developed to achieve high prediction accuracy, easy interpretation, and efficient computation simultaneously. The LIFE algorithm is able to fit a wide single-hidden-layer neural network (NN) accurately with three steps: defining the subsets of a dataset by the linear projections of neural nodes, creating the features from multiple narrow single-hidden-layer NNs trained on the different subsets of the data, combining the features with a linear model. The theoretical rationale behind LIFE is also provided by the connection to the loss ambiguity decomposition of stack ensemble methods. Both simulation and empirical experiments confirm that LIFE consistently outperforms directly trained single-hidden-layer NNs and also outperforms many other benchmark models, including multilayers feed forward neural network (FFNN), Xgboost, and random forest (RF) in many experiments. As a wide single-hidden-layer NN, LIFE is intrinsically interpretable. Meanwhile, both variable importance and global main and interaction effects can be easily created and visualized. In addition, the parallel nature of the base learner building makes LIFE computationally efficient by leveraging parallel computing.
引用
收藏
页码:9657 / 9685
页数:29
相关论文
共 50 条
  • [31] Accounting for model errors in iterative ensemble smoothers
    Geir Evensen
    Computational Geosciences, 2019, 23 : 761 - 775
  • [32] An efficient interpretable stacking ensemble model for lung cancer prognosis
    Arif, Umair
    Zhang, Chunxia
    Hussain, Sajid
    Abbasi, Abdul Rauf
    COMPUTATIONAL BIOLOGY AND CHEMISTRY, 2024, 113
  • [33] An interpretable deep feature aggregation framework for machinery incremental fault diagnosis
    Hu, Kui
    Chen, Qian
    Yao, Jintao
    He, Qingbo
    Peng, Zhike
    ADVANCED ENGINEERING INFORMATICS, 2025, 65
  • [34] Residual Sketch Learning for a Feature-Importance-Based and Linguistically Interpretable Ensemble Classifier
    Bian, Zekang
    Zhang, Jin
    Chung, Fu-Lai
    Wang, Shitong
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (08) : 10461 - 10474
  • [35] A Common Symmetrization Framework for Iterative (Linear) Maps
    Sarlette, Alain
    GEOMETRIC SCIENCE OF INFORMATION, GSI 2015, 2015, 9389 : 685 - 692
  • [36] Iterative ensemble feature selection for multiclass classification of imbalanced microarray data
    Yang, Junshan
    Zhou, Jiarui
    Zhu, Zexuan
    Ma, Xiaoliang
    Ji, Zhen
    JOURNAL OF BIOLOGICAL RESEARCH-THESSALONIKI, 2016, 23
  • [37] A two-sided iterative framework for model reduction of linear systems with quadratic output
    Gosea, Ion Victor
    Antoulas, Athanasios C.
    2019 IEEE 58TH CONFERENCE ON DECISION AND CONTROL (CDC), 2019, : 7812 - 7817
  • [38] A classification framework for multivariate compositional data with Dirichlet feature embedding
    Gu, Jie
    Cui, Bin
    Lu, Shan
    KNOWLEDGE-BASED SYSTEMS, 2021, 212
  • [39] An ensemble model for link prediction based on graph embedding
    Chen, Yen-Liang
    Hsiao, Chen-Hsin
    Wu, Chia-Chi
    DECISION SUPPORT SYSTEMS, 2022, 157
  • [40] Graph regularized locally linear embedding for unsupervised feature selection
    Miao, Jianyu
    Yang, Tiejun
    Sun, Lijun
    Fei, Xuan
    Niu, Lingfeng
    Shi, Yong
    PATTERN RECOGNITION, 2022, 122