论文标题
MDA:可用性意识到联合学习客户选择
MDA: Availability-Aware Federated Learning Client Selection
论文作者
论文摘要
最近,引入了一种新的分布式学习计划,称为联邦学习(FL)。 FL的设计使服务器永远不会收集用户拥有的数据,这意味着它可以保留隐私。 FL的过程始于服务器向客户端发送模型,然后客户使用其数据训练该模型,然后将更新的模型发送回服务器。之后,服务器汇总了所有更新并修改全局模型。重复此过程,直到模型收敛为止。这项研究的重点是称为Cross Device FL的FL设置,该设置根据大量客户进行训练。由于许多设备在跨设备FL中可能无法使用,并且服务器和所有客户之间的通信非常昂贵,因此只有一小部分客户在每个回合中都选择培训。在Vanilla FL中,随机选择客户,这会导致可接受的准确性,但从整体培训时间角度来看,这不是理想的选择,因为某些客户速度很慢,并且可能导致一些训练回合很慢。如果只有快速的客户被选中,学习就会加快,但它只会偏向快速客户的数据和准确性降低。因此,已经提出了新的客户选择技术来通过考虑各个客户的资源和速度来改善培训时间。本文介绍了第一个称为MDA的可用性选择策略。结果表明,我们的方法使比香草FL的学习速度高达6.5%。此外,我们表明资源异质性感知的技术是有效的,但在与我们的方法结合使用时可以变得更好,使其比最先进的选择者更快地提高了16%。最后,与仅选择快速客户端的客户选择器相比,我们的方法选择了更多独特的培训客户端,从而减少了我们的技术偏见。
Recently, a new distributed learning scheme called Federated Learning (FL) has been introduced. FL is designed so that server never collects user-owned data meaning it is great at preserving privacy. FL's process starts with the server sending a model to clients, then the clients train that model using their data and send the updated model back to the server. Afterward, the server aggregates all the updates and modifies the global model. This process is repeated until the model converges. This study focuses on an FL setting called cross-device FL, which trains based on a large number of clients. Since many devices may be unavailable in cross-device FL, and communication between the server and all clients is extremely costly, only a fraction of clients gets selected for training at each round. In vanilla FL, clients are selected randomly, which results in an acceptable accuracy but is not ideal from the overall training time perspective, since some clients are slow and can cause some training rounds to be slow. If only fast clients get selected the learning would speed up, but it will be biased toward only the fast clients' data, and the accuracy degrades. Consequently, new client selection techniques have been proposed to improve the training time by considering individual clients' resources and speed. This paper introduces the first availability-aware selection strategy called MDA. The results show that our approach makes learning faster than vanilla FL by up to 6.5%. Moreover, we show that resource heterogeneity-aware techniques are effective but can become even better when combined with our approach, making it faster than the state-of-the-art selectors by up to 16%. Lastly, our approach selects more unique clients for training compared to client selectors that only select fast clients, which reduces our technique's bias.