首页> 外文期刊>Computer Communications >Privacy preserving distributed machine learning with federated learning
【24h】

Privacy preserving distributed machine learning with federated learning

机译:利用联合学习保留分布式机器学习的隐私

获取原文
获取原文并翻译 | 示例
           

摘要

Edge computing and distributed machine learning have advanced to a level that can revolutionize a particular organization. Distributed devices such as the Internet of Things (IoT) often produce a large amount of data, eventually resulting in big data that can be vital in uncovering hidden patterns, and other insights in numerous fields such as healthcare, banking, and policing. Data related to areas such as healthcare and banking can contain potentially sensitive data that can become public if they are not appropriately sanitized. Federated learning (FedML) is a recently developed distributed machine learning (DML) approach that tries to preserve privacy by bringing the learning of an ML model to data owners' devices. However, literature shows different attack methods such as membership inference that exploit the vulnerabilities of ML models as well as the coordinating servers to retrieve private data. Hence, FedML needs additional measures to guarantee data privacy. Furthermore, big data often requires more resources than available in a standard computer. This paper addresses these issues by proposing a distributed perturbation algorithm named as DISTPAB, for privacy preservation of horizontally partitioned data. DISTPAB alleviates computational bottlenecks by distributing the task of privacy preservation utilizing the asymmetry of resources of a distributed environment, which can have resource-constrained devices as well as high-performance computers. Experiments show that DISTPAB provides high accuracy, high efficiency, high scalability, and high attack resistance. Further experiments on privacy-preserving FedML show that DISTPAB is an excellent solution to stop privacy leaks in DML while preserving high data utility.
机译:边缘计算和分布式机器学习已经进入了一个可以彻底改变特定组织的级别。诸如事物互联网(物联网)之类的分布式设备通常会产生大量数据,最终导致大数据在揭露隐藏模式中可能是至关重要的,以及诸如医疗保健,银行和警务等众多领域的其他见解。与医疗保健和银行等领域相关的数据可以包含可能成为公开的潜在敏感数据,如果他们没有适当消毒。联合学习(FEDML)是最近开发的分布式机器学习(DML)方法,它通过将ML模型的学习提供给数据所有者的设备来试图保护隐私。但是,文献显示了不同的攻击方法,例如隶属推断,用于利用ML模型的漏洞以及协调服务器来检索私有数据。因此,FEDML需要额外的措施来保证数据隐私。此外,大数据通常需要比标准计算机中的更多资源。本文通过提出名为DistPab的分布式扰动算法来解决这些问题,用于水平分区数据的隐私保存。 DistPab通过利用分布式环境的资源的不对称,通过分发隐私保存的任务来减轻计算瓶颈,这可以具有资源受限的设备以及高性能计算机。实验表明,Distpab提供了高精度,高效率,高可伸缩性和高攻击性。保留隐私保留FEDML的进一步实验显示,DistPab是在保留高数据实用程序时停止DML中的隐私泄漏的优秀解决方案。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号