论文标题

关于会员推理攻击的信誉

On the Discredibility of Membership Inference Attacks

论文作者

Rezaei, Shahbaz, Liu, Xin

论文摘要

随着机器学习模型的广泛应用,研究对敏感数据训练的模型的潜在数据泄漏已经变得至关重要。最近,提出了各种会员推理(MI)攻击,以确定样本是否是培训集的一部分。问题是这些攻击是否可以在实践中可靠地使用。我们表明,MI模型经常将成员样本的相邻非成员样本错误分类为成员。换句话说,它们对可以识别的确切成员样本的子群体具有很高的假阳性率。然后,我们展示了MI攻击的实际应用,在该问题中,此问题具有现实世界的影响。在这里,外部审核员(调查员)使用MI攻击向法官/陪审团展示审核员非法使用敏感数据。由于MI对成员亚群的高度误报率很高,Auditee通过揭示了MI攻击对这些亚群的表现,从而挑战了审计师的信誉。我们认为,当前的成员推理攻击可以识别记忆的亚群,但他们无法可靠地识别培训期间使用亚群中的哪个确切样本。

With the wide-spread application of machine learning models, it has become critical to study the potential data leakage of models trained on sensitive data. Recently, various membership inference (MI) attacks are proposed to determine if a sample was part of the training set or not. The question is whether these attacks can be reliably used in practice. We show that MI models frequently misclassify neighboring nonmember samples of a member sample as members. In other words, they have a high false positive rate on the subpopulations of the exact member samples that they can identify. We then showcase a practical application of MI attacks where this issue has a real-world repercussion. Here, MI attacks are used by an external auditor (investigator) to show to a judge/jury that an auditee unlawfully used sensitive data. Due to the high false positive rate of MI attacks on member's subpopulations, auditee challenges the credibility of the auditor by revealing the performance of the MI attacks on these subpopulations. We argue that current membership inference attacks can identify memorized subpopulations, but they cannot reliably identify which exact sample in the subpopulation was used during the training.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源