论文标题
深思熟虑吗?选择性对抗攻击,用于对多个深神经网络进行细粒度操纵
Do Deep Minds Think Alike? Selective Adversarial Attacks for Fine-Grained Manipulation of Multiple Deep Neural Networks
论文作者
论文摘要
最近的作品证明了针对单个机器学习系统的{\ it对抗性示例的存在。在本文中,我们提出了一个简单但基本的问题,即“选择性欺骗”:分配的{\ it多重}机器学习系统,用于解决相同的分类问题并采用相同的输入信号,是否有可能构建对输入信号的扰动,以操纵这些{\ IT多重}机器学习系统{\ IT-iT-int-fef fef-It-nitife precultane prefience in mortife prectience in mortife precultiate {例如,是否可以选择性地欺骗一组“敌方”机器学习系统,但不会欺骗其他“朋友”机器学习系统?这个问题的答案取决于这些不同的机器学习系统“相似”的程度。我们将“选择性欺骗”作为一个新的优化问题提出问题,并在MNIST数据集上报告一系列实验。我们从这些实验中的初步发现表明,即使分类器在其体系结构,培训算法和培训数据集相同的情况下,同时选择性地操纵多个MNIST分类器也很容易,除了在培训过程中随机初始化外。这表明,两个名义上等效的机器学习系统实际上根本不一样,并为许多新型应用程序和对深神经网络的工作原理的更深入理解打开了可能性。
Recent works have demonstrated the existence of {\it adversarial examples} targeting a single machine learning system. In this paper we ask a simple but fundamental question of "selective fooling": given {\it multiple} machine learning systems assigned to solve the same classification problem and taking the same input signal, is it possible to construct a perturbation to the input signal that manipulates the outputs of these {\it multiple} machine learning systems {\it simultaneously} in arbitrary pre-defined ways? For example, is it possible to selectively fool a set of "enemy" machine learning systems but does not fool the other "friend" machine learning systems? The answer to this question depends on the extent to which these different machine learning systems "think alike". We formulate the problem of "selective fooling" as a novel optimization problem, and report on a series of experiments on the MNIST dataset. Our preliminary findings from these experiments show that it is in fact very easy to selectively manipulate multiple MNIST classifiers simultaneously, even when the classifiers are identical in their architectures, training algorithms and training datasets except for random initialization during training. This suggests that two nominally equivalent machine learning systems do not in fact "think alike" at all, and opens the possibility for many novel applications and deeper understandings of the working principles of deep neural networks.