论文标题
安全加权聚合以用于联合学习
Secure Weighted Aggregation for Federated Learning
论文作者
论文摘要
互联网连接的数字服务的普遍采用导致了客户个人数据隐私的越来越关注。另一方面,数字服务提供商广泛采用了机器学习(ML)技术,以提高运营生产率和客户满意度。 ML不可避免地访问和处理用户的个人数据,如果不仔细执行,这可能会违反相关的隐私保护法规。当捕获用户数据并存储在分布式位置中时,基于云的数字服务的实现会加剧这种情况,因此ML的用户数据汇总可能是严重违反隐私法规的。在此背景下,联邦学习(FL)是一个新兴领域,允许在分布式数据上进行ML,而无需数据留下的位置。但是,根据数字服务的性质,在不同位置捕获的数据可能对业务运营具有不同的意义,因此,加权聚合对于增强了FLEARNEARNEAR模型的质量是非常可取的。此外,为防止用户数据从聚合梯度中泄漏,需要加密机制来允许FL的安全聚合。在本文中,我们提出了一种支持安全加权聚合的隐私增强佛罗里达计划。此外,通过设计基于零知识证明(ZKP)的验证协议,该计划的方案能够防止FL参与者的欺诈性信息。实验结果表明,我们的计划是实用且安全的。与现有的FL方法相比,我们的计划通过额外的安全保证来实现安全的加权汇总,以抵抗欺诈性消息,其负担得起的1.2倍运行时开销和通信成本的1.3倍。
The pervasive adoption of Internet-connected digital services has led to a growing concern in the personal data privacy of their customers. On the other hand, machine learning (ML) techniques have been widely adopted by digital service providers to improve operational productivity and customer satisfaction. ML inevitably accesses and processes users' personal data, which could potentially breach the relevant privacy protection regulations if not performed carefully. The situation is exacerbated by the cloud-based implementation of digital services when user data are captured and stored in distributed locations, hence aggregation of the user data for ML could be a serious breach of privacy regulations. In this backdrop, Federated Learning (FL) is an emerging area that allows ML on distributed data without the data leaving their stored location. However, depending on the nature of the digital services, data captured at different locations may carry different significance to the business operation, hence a weighted aggregation will be highly desirable for enhancing the quality of the FL-learned model. Furthermore, to prevent leakage of user data from the aggregated gradients, cryptographic mechanisms are needed to allow secure aggregation of FL. In this paper, we propose a privacy-enhanced FL scheme for supporting secure weighted aggregation. Besides, by devising a verification protocol based on Zero-Knowledge Proof (ZKP), the proposed scheme is capable of guarding against fraudulent messages from FL participants. Experimental results show that our scheme is practical and secure. Compared to existing FL approaches, our scheme achieves secure weighted aggregation with an additional security guarantee against fraudulent messages with an affordable 1.2 times runtime overheads and 1.3 times communication costs.