论文标题

捍卫回归学习者免于中毒攻击

Defending Regression Learners Against Poisoning Attacks

论文作者

Weerasinghe, Sandamal, Erfani, Sarah M., Alpcan, Tansu, Leckie, Christopher, Kopacz, Justin

论文摘要

从工程应用程序到财务预测广泛使用的回归模型容易受到针对性的恶意攻击,例如训练数据中毒,对手可以通过这些攻击来操纵其预测。试图解决此问题的先前作品取决于对攻击/攻击者性质的假设或高估了学习者的知识,使其不切实际。我们引入了一种新型的局部固有维度(LID)措施,称为N-LID,该测量衡量给定数据点盖子相对于其邻居的局部偏差。然后,我们表明N-LID可以将中毒样品与正常样品区分开,并提出一种基于N-LID的防御方法,该方法没有对攻击者的假设。通过使用基准数据集进行广泛的数值实验,我们表明,所提出的防御机制在预测准确性方面优于最先进的防御状态(与未防御的山脊模型相比,MSE降低了76%)和运行时间。

Regression models, which are widely used from engineering applications to financial forecasting, are vulnerable to targeted malicious attacks such as training data poisoning, through which adversaries can manipulate their predictions. Previous works that attempt to address this problem rely on assumptions about the nature of the attack/attacker or overestimate the knowledge of the learner, making them impractical. We introduce a novel Local Intrinsic Dimensionality (LID) based measure called N-LID that measures the local deviation of a given data point's LID with respect to its neighbors. We then show that N-LID can distinguish poisoned samples from normal samples and propose an N-LID based defense approach that makes no assumptions of the attacker. Through extensive numerical experiments with benchmark datasets, we show that the proposed defense mechanism outperforms the state of the art defenses in terms of prediction accuracy (up to 76% lower MSE compared to an undefended ridge model) and running time.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源