首页> 外文期刊>International Journal of Intelligent Systems >PrivateDL: Privacy-preserving collaborative deep learning against leakage from gradient sharing
【24h】

PrivateDL: Privacy-preserving collaborative deep learning against leakage from gradient sharing

机译:私有化:保护渐变分享的泄漏的隐私合作深度学习

获取原文
获取原文并翻译 | 示例
           

摘要

Large-scale data training is vital to the generalization performance of deep learning (DL) models. However, collecting data directly is associated with increased risk of privacy disclosure, particularly in special fields such as healthcare, finance, and genomics. To protect training data privacy, collaborative deep learning (CDL) has been proposed to enable joint training from multiple data owners while providing reliable privacy guarantee. However, recent studies have shown that CDL is vulnerable to several attacks that could reveal sensitive information about the original training data. One of the most powerful attacks benefits from the leakage from gradient sharing during collaborative training process. In this study, we present a new CDL framework, PrivateDL, to effectively protect private training data against leakage from gradient sharing. Unlike conventional training process that trains on private data directly, PrivateDL allows effective transfer of relational knowledge from sensitive data to public data in a privacy-preserving way, and enables participants to jointly learn local models based on the public data with noise-preserving labels. This way, PrivateDL establishes a privacy gap between the local models and the private datasets, thereby ensuring privacy against the attacks launched to the local models through gradient sharing. Moreover, we propose a new algorithm called Distributed Aggregation Stochastic Gradient Descent, which is designed to improve the efficiency and accuracy of CDL, especially in the asynchronous training mode. Experimental results demonstrate that PrivateDL preserves data privacy with reasonable performance overhead.
机译:大规模数据培训对深度学习(DL)模型的泛化性能至关重要。然而,直接收集数据与隐私公开的风险增加相关,特别是在医疗保健,金融和基因组学等特殊领域。为了保护培训数据隐私,提议协作深度学习(CDL)能够从多个数据所有者开始培训,同时提供可靠的隐私保障。然而,最近的研究表明,CDL容易受到几种攻击,可以揭示有关原始培训数据的敏感信息。在协作培训过程中,来自梯度共享的泄漏中最强大的攻击之一。在这项研究中,我们提出了一个新的CDL框架,私有线,以有效保护私人培训数据免受梯度共享的泄漏。与直接私人数据进行列车的传统培训过程不同,私有化允许以隐私的方式从敏感数据从敏感数据转移到公共数据,并使参与者能够根据具有噪声保存标签的公共数据共同学习本地模型。这样,PrivationL在本地模型和私有数据集之间建立隐私差距,从而确保通过梯度共享将攻击攻击攻击到本地模型。此外,我们提出了一种称为分布式聚集随机梯度下降的新算法,旨在提高CDL的效率和准确性,尤其是在异步训练模式中。实验结果表明,私人可以通过合理的性能开销保留数据隐私。

著录项

  • 来源
    《International Journal of Intelligent Systems》 |2020年第8期|1262-1279|共18页
  • 作者单位

    School of Information Science and Engineering University of Jinan Jinan China Shandong Provincial Key Laboratory of Network-based Intelligent Computing University of Jinan Jinan China;

    School of Information Science and Engineering University of Jinan Jinan China Shandong Provincial Key Laboratory of Network-based Intelligent Computing University of Jinan Jinan China Shandong Provincial Key Laboratory of Software Engineering Jinan China;

    Imperial College London London UK;

    School of Information Science and Engineering University of Jinan Jinan China Shandong Provincial Key Laboratory of Network-based Intelligent Computing University of Jinan Jinan China;

    School of Information Science and Engineering University of Jinan Jinan China Shandong Provincial Key Laboratory of Network-based Intelligent Computing University of Jinan Jinan China;

  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    collaborative deep learning; gradient sharing; machine learning; privacy-preserving technique;

    机译:合作深度学习;梯度分享;机器学习;保护技术;

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号