论文标题
使用智能合同系统进行银行贷款预测的分散汇总机制,用于培训深度学习模型
A decentralized aggregation mechanism for training deep learning models using smart contract system for bank loan prediction
论文作者
论文摘要
在试图构建复杂的基于深度学习的系统以建模数据时,数据隐私和共享一直是一个关键问题。促进分散的方法可以从多个节点跨数据中受益的同时不需要在物理上合并数据的数据是一个积极研究的领域。在本文中,我们提出了一种解决方案,以通过使用智能合同系统来培训深度学习体系结构,从分布式数据设置中受益。具体而言,我们提出了一种机制,该机制将从区块链上从本地ANN模型获得的中间表示汇总在一起。对本地模型的培训进行了各自的数据。将其组合和训练在主机节点上时,从它们得出的中间表示有助于获得更准确的系统。尽管联合学习主要处理在多个节点上分布的样本数量的相同功能,但在这里我们正在处理相同数量的样本,但它们的功能分布在多个节点上。我们考虑银行贷款预测的任务,其中个人及其特定于银行的详细信息可能在同一地点不可用。我们的聚合机制有助于对现有的分布式数据进行培训,而无需共享和加入实际数据值。获得的性能,比单个节点的性能要好,并且与集中式数据设置的性能相提并论,这是将我们的技术扩展到其他架构和任务中的有力基础。该解决方案在想要在垂直分区的数据上训练深度学习模型的组织中找到了应用程序。
Data privacy and sharing has always been a critical issue when trying to build complex deep learning-based systems to model data. Facilitation of a decentralized approach that could take benefit from data across multiple nodes while not needing to merge their data contents physically has been an area of active research. In this paper, we present a solution to benefit from a distributed data setup in the case of training deep learning architectures by making use of a smart contract system. Specifically, we propose a mechanism that aggregates together the intermediate representations obtained from local ANN models over a blockchain. Training of local models takes place on their respective data. The intermediate representations derived from them, when combined and trained together on the host node, helps to get a more accurate system. While federated learning primarily deals with the same features of data where the number of samples being distributed on multiple nodes, here we are dealing with the same number of samples but with their features being distributed on multiple nodes. We consider the task of bank loan prediction wherein the personal details of an individual and their bank-specific details may not be available at the same place. Our aggregation mechanism helps to train a model on such existing distributed data without having to share and concatenate together the actual data values. The obtained performance, which is better than that of individual nodes, and is at par with that of a centralized data setup makes a strong case for extending our technique across other architectures and tasks. The solution finds its application in organizations that want to train deep learning models on vertically partitioned data.