高级检索
    赵小阳, 李仲年, 王文玉, 许新征. ADIC:一种面向可解释图像识别的自适应解纠缠CNN分类器[J]. 计算机研究与发展, 2023, 60(8): 1754-1767. DOI: 10.7544/issn1000-1239.202330231
    引用本文: 赵小阳, 李仲年, 王文玉, 许新征. ADIC:一种面向可解释图像识别的自适应解纠缠CNN分类器[J]. 计算机研究与发展, 2023, 60(8): 1754-1767. DOI: 10.7544/issn1000-1239.202330231
    Zhao Xiaoyang, Li Zhongnian, Wang Wenyu, Xu Xinzheng. ADIC: An Adaptive Disentangled CNN Classifier for Interpretable Image Recognition[J]. Journal of Computer Research and Development, 2023, 60(8): 1754-1767. DOI: 10.7544/issn1000-1239.202330231
    Citation: Zhao Xiaoyang, Li Zhongnian, Wang Wenyu, Xu Xinzheng. ADIC: An Adaptive Disentangled CNN Classifier for Interpretable Image Recognition[J]. Journal of Computer Research and Development, 2023, 60(8): 1754-1767. DOI: 10.7544/issn1000-1239.202330231

    ADIC:一种面向可解释图像识别的自适应解纠缠CNN分类器

    ADIC: An Adaptive Disentangled CNN Classifier for Interpretable Image Recognition

    • 摘要: 近年来,卷积神经网络(convolutional neural network,CNN)作为一种典型的深度神经网络模型,在图像识别、目标检测和语义分割等计算机视觉领域中取得了令人瞩目的成效. 然而,CNN端到端的学习模式使其隐藏层的逻辑关系以及模型决策结果难以被解释,这限制了其推广应用. 因此,研究可解释的CNN具有重要意义和应用价值. 为了使CNN的分类器具有可解释性,近年来涌现出了很多在CNN架构中引入基础概念作为插入式成分的研究. 事后概念激活向量方法以基础概念为表现形式,用于分析预训练的模型,但依赖独立于原始模型的额外的分类器,解释结果可能并不符合原始模型逻辑. 另外,现有的一些基于概念的事前可解释方法对于CNN潜在分类空间中的概念处理太过绝对. 引入图卷积网络模块,设计了一种类内概念图编码器(within-class concepts graphs encoder, CGE)学习类内基础概念及其潜在交互. 在CGE基础上,设计实现不同依赖关系的基础概念不同程度解纠缠的正则化项,提出了潜在空间自适应解纠缠的可解释CNN分类器(adaptive disentangled interpretable CNN classifier, ADIC). 将ADIC嵌入ResNet-18和ResNet-50架构,在Mini-ImageNet和Places365数据集上的分类实验和可解释图像识别实验结果表明,ADIC在保证基准模型具有自解释能力的前提下,可以进一步提高基准模型的精度.

       

      Abstract: In recent years, convolutional neural network (CNN), as a typical deep neural network model, has achieved remarkable results in computer vision fields such as image recognition, target detection and semantic segmentation. However, the end-to-end learning mode of CNNs makes the logical relationships of their hidden layers and the results of model decisions difficult to be interpreted, which limits their promotion and application. Therefore, the research of interpretable CNNs is of important significance and application value. In order to make the classifier of CNNs interpretable, many researches have emerged in recent years to introduce basis concepts into CNN architectures as plug-in components. The post-hoc concept activation vector methods take the basis concept as their representation and are used to analyze the pre-trained models. However, they rely on additional classifiers independent of the original models and the interpretation results may not match the original model logic. Furthermore, some existing concept-based ad-hoc interpretable methods are too absolute in handling concepts in the latent classification space of CNNs. In this work, a within-class concepts graphs encoder (CGE) is designed by introducing a graph convolutional network module to learn the basis concepts within a class and their latent interactions. The adaptive disentangled interpretable CNN classifier (ADIC) with adaptive disentangled latent space is proposed based on CGE by designing regularization terms that implement different degrees disentanglement of the basis concepts with different dependencies. By embedding ADIC into ResNet18 and ResNet50 architectures, classification experiments and interpretable image recognition experiments on Mini-ImageNet and Places 365 datasets have shown that ADIC can further improve the accuracy of the baseline model while ensuring that the baseline model has self-interpretability.

       

    /

    返回文章
    返回