论文标题

通过一致性正规化来减轻对抗训练的强大过度拟合

Alleviating Robust Overfitting of Adversarial Training With Consistency Regularization

论文作者

Zhang, Shudong, Gao, Haichang, Zhang, Tianwei, Zhou, Yunyi, Wu, Zihui

论文摘要

事实证明,对抗性训练(AT)是捍卫深层神经网络(DNN)免受对抗攻击的最有效方法之一。但是,强大过度拟合的现象,即,鲁棒性在某个阶段将急剧下降,始终存在于AT期间。为了获得强大的模型,减少这种稳健的概括差距非常重要。在本文中,我们提出了一项深入的研究,旨在从新角度出发进行强大的过度拟合。我们观察到,一致性正则化是一种半监督学习中的一种流行技术,其目标与AT相似,可以用来减轻强大的过度拟合。我们从经验上验证了这一观察结果,发现大多数先前的解决方案都与一致性正规化具有隐式联系。在此激励的情况下,我们引入了一种新的解决方案,该解决方案将一致性正则化和卑鄙的老师(MT)策略整合到AT中。具体来说,我们介绍了一个教师模型,该模型来自培训步骤中学生模型的平均权重。然后,我们设计一个一致性损失函数,以使学生模型的预测分布在对抗性示例上,这与老师模型在干净的样本中的预测分布。实验表明,我们提出的方法可以有效地减轻强大的过度拟合,并改善DNN模型对常见的对抗攻击的鲁棒性。

Adversarial training (AT) has proven to be one of the most effective ways to defend Deep Neural Networks (DNNs) against adversarial attacks. However, the phenomenon of robust overfitting, i.e., the robustness will drop sharply at a certain stage, always exists during AT. It is of great importance to decrease this robust generalization gap in order to obtain a robust model. In this paper, we present an in-depth study towards the robust overfitting from a new angle. We observe that consistency regularization, a popular technique in semi-supervised learning, has a similar goal as AT and can be used to alleviate robust overfitting. We empirically validate this observation, and find a majority of prior solutions have implicit connections to consistency regularization. Motivated by this, we introduce a new AT solution, which integrates the consistency regularization and Mean Teacher (MT) strategy into AT. Specifically, we introduce a teacher model, coming from the average weights of the student models over the training steps. Then we design a consistency loss function to make the prediction distribution of the student models over adversarial examples consistent with that of the teacher model over clean samples. Experiments show that our proposed method can effectively alleviate robust overfitting and improve the robustness of DNN models against common adversarial attacks.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源