ADIC: An Adaptive Disentangled CNN Classifier for Interpretable Image Recognition

被引:0
|
作者
Zhao X. [1 ]
Li Z. [1 ]
Wang W. [1 ]
Xu X. [1 ,2 ]
机构
[1] School of Computer Science and Technology, China University of Mining and Technology, Jiangsu, Xuzhou
[2] Engineering Research Center of Digital mine, China University of Mining and Technology, Ministry of Education, Jiangsu, Xuzhou
基金
中国国家自然科学基金;
关键词
category basis concept; convolutional neural network; disentangle; graph convolution network; interpretability;
D O I
10.7544/issn1000-1239.202330231
中图分类号
学科分类号
摘要
In recent years, convolutional neural network (CNN), as a typical deep neural network model, has achieved remarkable results in computer vision fields such as image recognition, target detection and semantic segmentation. However, the end-to-end learning mode of CNNs makes the logical relationships of their hidden layers and the results of model decisions difficult to be interpreted, which limits their promotion and application. Therefore, the research of interpretable CNNs is of important significance and application value. In order to make the classifier of CNNs interpretable, many researches have emerged in recent years to introduce basis concepts into CNN architectures as plug-in components. The post-hoc concept activation vector methods take the basis concept as their representation and are used to analyze the pre-trained models. However, they rely on additional classifiers independent of the original models and the interpretation results may not match the original model logic. Furthermore, some existing concept-based ad-hoc interpretable methods are too absolute in handling concepts in the latent classification space of CNNs. In this work, a within-class concepts graphs encoder (CGE) is designed by introducing a graph convolutional network module to learn the basis concepts within a class and their latent interactions. The adaptive disentangled interpretable CNN classifier (ADIC) with adaptive disentangled latent space is proposed based on CGE by designing regularization terms that implement different degrees disentanglement of the basis concepts with different dependencies. By embedding ADIC into ResNet18 and ResNet50 architectures, classification experiments and interpretable image recognition experiments on Mini-ImageNet and Places 365 datasets have shown that ADIC can further improve the accuracy of the baseline model while ensuring that the baseline model has self-interpretability. © 2023 Science Press. All rights reserved.
引用
收藏
页码:1754 / 1767
页数:13
相关论文
共 44 条
  • [1] Been K, Martin W, Justin G, Et al., Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (TCAV)[C], Proc of the 35th Int Conf on Machine Learning (ICML), pp. 2668-2677, (2018)
  • [2] Amirata G, James W, James Y Z, Et al., Towards automatic concept-based explanations[C], Proc of the Conf on Advances in Neural Information Processing Systems (NeurIPS), pp. 9273-9282, (2019)
  • [3] Zhang Ruihan, Prashan M, Tim M, Et al., Invertible concept-based explanations for CNN models with non-negative concept activation vectors[C], Proc of the AAAI Conf on Artificial Intelligence, pp. 11682-11690, (2021)
  • [4] Zhi Chen, Yijie Bei, Cynthia R., Concept whitening for interpretable image recognition[J], Nature Machine Intelligence, 2, 12, pp. 772-782, (2020)
  • [5] Wang Jiaqi, Liu Huafeng, Wang Xinyue, Et al., Interpretable image recognition by constructing transparent embedding space[C], Proc of the IEEE Int Conf on Computer Vision (ICCV), pp. 875-884, (2021)
  • [6] Jon D, Alina J B, Chen Chaofan, Deformable ProtoPNet: An interpretable image classifier using deformable prototypes, Proc of the IEEE/CVF Conf on Computer Vision and Pattern Recognition (CVPR), pp. 10255-10265, (2022)
  • [7] Francesco B, Fosca G, Riccardo G, Et al., Benchmarking and survey of explanation methods for black box models, (2021)
  • [8] Shouling Ji, Jinfeng Li, Tianyu Du, Et al., A survey of interpretability methods, applications and security of machine learning models[J], Journal of Computer Research and Development, 56, 10, pp. 2071-2096, (2019)
  • [9] Pengbo Yang, Jitao Sang, Biao Zhang, Et al., Survey of the interpretability of deep models for image classification[J], Journal of Software, 34, 1, pp. 230-254, (2023)
  • [10] Chatonsky G., Deep dream (The Network's Dream)[J], SubStance, 45, 2, pp. 61-77, (2016)