首页> 外文会议>IEEE Statistical Signal Processing Workshop >Differential Privacy for Positive and Unlabeled Learning With Known Class Priors
【24h】

Differential Privacy for Positive and Unlabeled Learning With Known Class Priors

机译:具有已知课程先验的积极和无标签学习的差异隐私

获取原文

摘要

Despite the increasing attention to big data, there are several domains where labeled data is scarce or too costly to obtain. For example, for data from information retrieval, gene analysis, and social network analysis, only training samples from the positive class are annotated while the remaining unlabeled training samples consist of both unlabeled positive and unlabeled negative samples. The specific positive and unlabeled (PU) data from those domains necessitates a mechanism to learn a two-class classifier from only one-class labeled data. Moreover, because data from those domains is highly sensitive and private, preserving training samples privacy is essential. This paper addresses the challenge of private PU learning by designing a differentially private algorithm for positive and unlabeled data. We first propose a learning framework for the PU setting when the class prior probability is known, with a theoretical guarantee of convergence to the optimal classifier. We then propose a privacy-preserving mechanism for the designed framework where the privacy and utility are both theoretically and empirically proved.
机译:尽管人们越来越关注大数据,但在某些领域中,标记数据却很少或获取成本太高。例如,对于来自信息检索,基因分析和社交网络分析的数据,仅注释来自阳性类别的训练样本,而其余未标记的训练样本由未标记的阳性和未标记的阴性样本组成。来自这些域的特定正数和未标记(PU)数据需要一种仅从一类标记数据中学习两类分类器的机制。此外,由于来自这些域的数据高度敏感且私有,因此保留培训样本的隐私至关重要。本文通过针对阳性和未标记数据设计差分私有算法来解决私有PU学习的挑战。当类优先级概率已知时,我们首先为PU设置提出一个学习框架,从理论上保证可以收敛到最优分类器。然后,我们为设计的框架提出了一种隐私保护机制,其中在理论上和经验上都证明了隐私和实用性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号