论文标题
解开可解释图像识别的深度学习
Decoupling Deep Learning for Interpretable Image Recognition
论文作者
论文摘要
神经网络的可解释性最近受到了广泛的关注。以前的基于原型的可解释网络涉及推理和解释过程中的原型激活,这需要针对原型的特定可解释的结构,从而使网络随着可解释性而降低准确。因此,提出了脱钩的原型网络(dprotonet),以避免此问题。这个新模型包含编码器,推理和解释模块。关于编码器模块,提出了不受限制的特征掩模,以生成表达性特征和原型。关于推理模块,引入了多图像原型学习方法以更新原型,以便网络可以学习广义原型。最后,关于解释模块,建议使用多个动态掩码(MDM)解码器来解释神经网络,该神经网络在网络的检测节点上使用原始图像和掩码图像的一致激活生成热图。它通过避免使用原型激活来解释网络的决策,以同时提高神经网络的准确性和解释性来解释基于原型网络的推理和解释模块。测试了多个公共通用和医疗数据集,结果证实,与以前的方法相比,我们的方法可以提高准确性和最先进的解释性。
The interpretability of neural networks has recently received extensive attention. Previous prototype-based explainable networks involved prototype activation in both reasoning and interpretation processes, requiring specific explainable structures for the prototype, thus making the network less accurate as it gains interpretability. Therefore, the decoupling prototypical network (DProtoNet) was proposed to avoid this problem. This new model contains encoder, inference, and interpretation modules. As regards the encoder module, unrestricted feature masks were presented to generate expressive features and prototypes. Regarding the inference module, a multi-image prototype learning method was introduced to update prototypes so that the network can learn generalized prototypes. Finally, concerning the interpretation module, a multiple dynamic masks (MDM) decoder was suggested to explain the neural network, which generates heatmaps using the consistent activation of the original image and mask image at the detection nodes of the network. It decouples the inference and interpretation modules of a prototype-based network by avoiding the use of prototype activation to explain the network's decisions in order to simultaneously improve the accuracy and interpretability of the neural network. The multiple public general and medical datasets were tested, and the results confirmed that our method could achieve a 5% improvement in accuracy and state-of-the-art interpretability compared with previous methods.