首页> 外文会议>IEEE Symposium on Security and Privacy >Comprehensive Privacy Analysis of Deep Learning: Passive and Active White-box Inference Attacks against Centralized and Federated Learning
【24h】

Comprehensive Privacy Analysis of Deep Learning: Passive and Active White-box Inference Attacks against Centralized and Federated Learning

机译:深度学习的全面隐私分析:针对集中式和联合学习的被动式和主动式白盒推理攻击

获取原文

摘要

Deep neural networks are susceptible to various inference attacks as they remember information about their training data. We design white-box inference attacks to perform a comprehensive privacy analysis of deep learning models. We measure the privacy leakage through parameters of fully trained models as well as the parameter updates of models during training. We design inference algorithms for both centralized and federated learning, with respect to passive and active inference attackers, and assuming different adversary prior knowledge. We evaluate our novel white-box membership inference attacks against deep learning algorithms to trace their training data records. We show that a straightforward extension of the known black-box attacks to the white-box setting (through analyzing the outputs of activation functions) is ineffective. We therefore design new algorithms tailored to the white-box setting by exploiting the privacy vulnerabilities of the stochastic gradient descent algorithm, which is the algorithm used to train deep neural networks. We investigate the reasons why deep learning models may leak information about their training data. We then show that even well-generalized models are significantly susceptible to white-box membership inference attacks, by analyzing state-of-the-art pre-trained and publicly available models for the CIFAR dataset. We also show how adversarial participants, in the federated learning setting, can successfully run active membership inference attacks against other participants, even when the global model achieves high prediction accuracies.
机译:深度神经网络会记住有关其训练数据的信息,因此容易受到各种推理攻击。我们设计白盒推理攻击来对深度学习模型进行全面的隐私分析。我们通过训练有素的模型的参数以及训练过程中模型的参数更新来衡量隐私泄漏。我们针对被动和主动推理攻击者设计了用于集中式学习和联合学习的推理算法,并假设了不同的对手先验知识。我们评估了针对深度学习算法的新型白盒成员推断攻击,以追踪其训练数据记录。我们表明,将已知的黑盒攻击直接扩展到白盒设置(通过分析激活函数的输出)是无效的。因此,我们通过利用随机梯度下降算法(用于训练深度神经网络的算法)的隐私漏洞,设计了适合白盒设置的新算法。我们调查了深度学习模型可能泄漏其训练数据信息的原因。然后,我们通过分析CIFAR数据集的最新技术,经过预先训练和可公开获得的模型,来证明即使是通用化的模型也极易受到白盒成员推断攻击的影响。我们还展示了在联合学习环境中,对抗性参与者如何能够成功地对其他参与者进行主动成员推理攻击,即使全局模型实现了较高的预测精度也是如此。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号