Re-interpreting rules interpretability

被引:0
|
作者
Adilova, Linara [1 ,2 ]
Kamp, Michael [1 ,3 ,4 ]
Andrienko, Gennady [2 ,5 ]
Andrienko, Natalia [2 ,5 ]
机构
[1] Ruhr Univ Bochum, D-44801 Bochum, Germany
[2] Fraunhofer Inst IAIS, D-53757 St Augustin, Germany
[3] Univ Med Essen, IKIM, D-45131 Essen, Germany
[4] Monash Univ, Melbourne, Vic 3800, Australia
[5] City Univ London, London EC1V 0HB, England
关键词
Interpretability; Descriptive model; Global explanation; Generalization; DECISION; SETS; REDUCTION; MODEL;
D O I
10.1007/s41060-023-00398-5
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Trustworthy machine learning requires a high level of interpretability of machine learning models, yet many models are inherently black-boxes. Training interpretable models instead-or using them to mimic the black-box model-seems like a viable solution. In practice, however, these interpretable models are still unintelligible due to their size and complexity. In this paper, we present an approach to explain the logic of large interpretable models that can be represented as sets of logical rules by a simple, and thus intelligible, descriptive model. The coarseness of this descriptive model and its fidelity to the original model can be controlled, so that a user can understand the original model in varying levels of depth. We showcase and discuss this approach on three real-world problems from healthcare, material science, and finance.
引用
收藏
页数:21
相关论文
共 50 条