论文标题
您看到的是您的分类:黑匣子归因
What You See is What You Classify: Black Box Attributions
论文作者
论文摘要
解释深图像分类器的重要一步在于识别图像区域,这些区域有助于模型输出中的单个班级分数。但是,由于此类网络的黑框性质,准确地进行此操作是一项艰巨的任务。大多数现有方法使用激活和梯度找到此类属性,或者通过反复扰动输入。相反,我们通过培训第二个深层网络The Expracter来应对这一挑战,以预测预先训练的黑盒分类器explanandum的归因。这些归因是以掩模的形式提供的,该掩码仅显示图像的相关部分,掩盖其余部分。与其他方法产生的显着图相比,我们的方法会产生更清晰和更具边界的掩盖。此外,与大多数现有方法不同,我们的方法能够在单个前向传球中直接生成非常不同的类别掩码。这使得在推断过程中提出的方法非常有效。我们表明,相对于Pascal VOC-2007和Microsoft Coco-2014数据集,我们的归因优于视觉和定量的既定方法。
An important step towards explaining deep image classifiers lies in the identification of image regions that contribute to individual class scores in the model's output. However, doing this accurately is a difficult task due to the black-box nature of such networks. Most existing approaches find such attributions either using activations and gradients or by repeatedly perturbing the input. We instead address this challenge by training a second deep network, the Explainer, to predict attributions for a pre-trained black-box classifier, the Explanandum. These attributions are provided in the form of masks that only show the classifier-relevant parts of an image, masking out the rest. Our approach produces sharper and more boundary-precise masks when compared to the saliency maps generated by other methods. Moreover, unlike most existing approaches, ours is capable of directly generating very distinct class-specific masks in a single forward pass. This makes the proposed method very efficient during inference. We show that our attributions are superior to established methods both visually and quantitatively with respect to the PASCAL VOC-2007 and Microsoft COCO-2014 datasets.