论文标题
如果可以的话,请找到它:端到端的对抗性擦除弱监督语义分段
Find it if You Can: End-to-End Adversarial Erasing for Weakly-Supervised Semantic Segmentation
论文作者
论文摘要
语义细分是一项传统上需要大量像素级地面真相标签的任务,这既耗时又昂贵。弱监督的设置中的最新进展表明,只能使用图像级标签获得合理的性能。分类通常被用作训练深度神经网络的代理任务,从中提取了注意力图。但是,分类任务只需要最低限度的证据来做出预测,因此它集中于最歧视的对象区域。为了克服这个问题,我们提出了一种新颖的对抗图图的对抗性擦除的表述。与以前的对抗性擦除方法相反,我们优化了两个具有相反损失功能的网络,从而消除了某些次优策略的要求;例如,有多个培训步骤使培训过程复杂化或在不同分布上运行的网络之间的重量共享策略,这可能是绩效的最佳选择。提出的解决方案不需要显着性掩码,而是使用正则化损失来防止注意图扩散到较小的歧视对象区域。我们在Pascal VOC数据集上的实验表明,与以前的对抗性擦除方法相比,与基线相比,我们的对抗方法将分割性能提高2.1 miOU,而我们的基线方法则增加了1.0 miou。
Semantic segmentation is a task that traditionally requires a large dataset of pixel-level ground truth labels, which is time-consuming and expensive to obtain. Recent advancements in the weakly-supervised setting show that reasonable performance can be obtained by using only image-level labels. Classification is often used as a proxy task to train a deep neural network from which attention maps are extracted. However, the classification task needs only the minimum evidence to make predictions, hence it focuses on the most discriminative object regions. To overcome this problem, we propose a novel formulation of adversarial erasing of the attention maps. In contrast to previous adversarial erasing methods, we optimize two networks with opposing loss functions, which eliminates the requirement of certain suboptimal strategies; for instance, having multiple training steps that complicate the training process or a weight sharing policy between networks operating on different distributions that might be suboptimal for performance. The proposed solution does not require saliency masks, instead it uses a regularization loss to prevent the attention maps from spreading to less discriminative object regions. Our experiments on the Pascal VOC dataset demonstrate that our adversarial approach increases segmentation performance by 2.1 mIoU compared to our baseline and by 1.0 mIoU compared to previous adversarial erasing approaches.