首页> 外文会议>IEEE Conference on Computer Communications >Optimizing Federated Learning on Non-IID Data with Reinforcement Learning
【24h】

Optimizing Federated Learning on Non-IID Data with Reinforcement Learning

机译:用加固学习优化非IID数据的联合学习

获取原文

摘要

The widespread deployment of machine learning applications in ubiquitous environments has sparked interests in exploiting the vast amount of data stored on mobile devices. To preserve data privacy, Federated Learning has been proposed to learn a shared model by performing distributed training locally on participating devices and aggregating the local models into a global one. However, due to the limited network connectivity of mobile devices, it is not practical for federated learning to perform model updates and aggregation on all participating devices in parallel. Besides, data samples across all devices are usually not independent and identically distributed (IID), posing additional challenges to the convergence and speed of federated learning. In this paper, we propose Favor, an experience-driven control framework that intelligently chooses the client devices to participate in each round of federated learning to counterbalance the bias introduced by non-IID data and to speed up convergence. Through both empirical and mathematical analysis, we observe an implicit connection between the distribution of training data on a device and the model weights trained based on those data, which enables us to profile the data distribution on that device based on its uploaded model weights. We then propose a mechanism based on deep Q-learning that learns to select a subset of devices in each communication round to maximize a reward that encourages the increase of validation accuracy and penalizes the use of more communication rounds. With extensive experiments performed in PyTorch, we show that the number of communication rounds required in federated learning can be reduced by up to 49% on the MNIST dataset, 23% on FashionMNIST, and 42% on CIFAR-10, as compared to the Federated Averaging algorithm.
机译:无处不在环境中的机器学习应用的广泛部署引发了利用存储在移动设备上的大量数据的兴趣。为了保留数据隐私,已经提出联合学习通过在参与设备上执行本地的分布式培训并将本地模型聚合到全局之一来学习共享模型。然而,由于移动设备的网络连接有限,因此联合学习是不实际的,用于并行地对所有参与设备执行模型更新和聚合。此外,所有设备上的数据样本通常都不是无关和相同分布的(IID),对联合学习的收敛和速度构成了额外的挑战。在本文中,我们提出了努力,经验驱动的控制框架,智能地选择客户端设备参与每一轮联合学习,以平衡非IID数据引入的偏差并加速收敛。通过经验和数学分析,我们观察设备上训练数据的分布与基于这些数据的模型权重之间的隐式连接,这使我们能够根据其上传的模型权重来配置该设备上的数据分布。然后,我们提出了一种基于Deep Q学习的机制,该机制学习在每个通信轮中选择一个设备的子集,以最大化鼓励验证准确性提高并惩罚更多通信轮的奖励。在Pytorch中进行了广泛的实验,我们表明联合学习所需的通信轮数量在Mnist DataSet上可以减少高达49%,与Federated相比,CIFAR-10的23%,42%平均算法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号