首页> 外文会议>Annual IFIP WG 11.3 conference on data and applications security and privacy >Practical Differentially Private Modeling of Human Movement Data
【24h】

Practical Differentially Private Modeling of Human Movement Data

机译:人体运动数据的实用差分私人建模

获取原文

摘要

Exciting advances in big data analysis suggest that sharing personal information, such as health and location data, among multiple other parties could have significant societal benefits. However, privacy issues often hinder data sharing. Recently, differential privacy emerged as an important tool to preserve privacy while sharing privacy-sensitive data. The basic idea is simple. Differential privacy guarantees that results learned from shared data do not change much based on the inclusion or exclusion of any single person's data. Despite the promise, existing differential privacy techniques addresses specific utility goals and/or query types (e.g., count queries), so it is not clear whether they can preserve utility for arbitrary types of queries. To better understand possible utility and privacy tradeoffs using differential privacy, we examined uses of human mobility data in a real-world competition. Participants were asked to come up with insightful ideas that leveraged a minimally protected published dataset. An obvious question is whether contest submissions could yield the same results if performed on a dataset protected by differential privacy? To answer this question, we studied synthetic dataset generation models for human mobility data using differential privacy. We discuss utility evaluation and the generality of the models extensively. Finally, we analyzed whether the proposed differential privacy models could be used in practice by examining contest submissions. Our results indicate that most of the competition submissions could be replicated using differentially private data with nearly the same utility and with privacy guarantees. Statistical comparisons with the original dataset demonstrate that differentially private synthetic versions of human mobility data can be widely applicable for data analysis.
机译:大数据分析中令人振奋的进步表明,在多个其他方之间共享个人信息(例如健康和位置数据)可能会带来巨大的社会效益。但是,隐私问题通常会阻碍数据共享。近来,差异隐私成为一种重要的工具,可以在共享敏感数据的同时保护隐私。基本思想很简单。差异性隐私保证了从共享数据中学到的结果不会因包含或排除任何个人数据而发生太大变化。尽管有希望,但是现有的差异隐私技术解决了特定的实用程序目标和/或查询类型(例如,计数查询),因此尚不清楚它们是否可以为任意类型的查询保留实用程序。为了更好地理解使用差异性隐私的可能的效用和隐私权衡,我们研究了在现实世界中竞赛中人类移动性数据的使用。要求参与者提出有洞察力的想法,这些想法利用受保护程度最低的已发布数据集。一个明显的问题是,如果在受差异性隐私保护的数据集上进行比赛,提交的参赛作品是否会产生相同的结果?为了回答这个问题,我们研究了使用差分隐私的人类移动性数据的综合数据集生成模型。我们将广泛讨论效用评估和模型的一般性。最后,我们通过检查比赛提交的内容,分析了提议的差异隐私模型是否可以在实践中使用。我们的结果表明,大多数竞赛提交的内容都可以使用具有几乎相同效用和隐私保证的差异私有数据来复制。与原始数据集的统计比较表明,人类流动性数据的不同私有合成版本可广泛应用于数据分析。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号