...
首页> 外文期刊>SIAM Journal on Computing >SAMPLE COMPLEXITY BOUNDS ON DIFFERENTIALLY PRIVATE LEARNING VIA COMMUNICATION COMPLEXITY
【24h】

SAMPLE COMPLEXITY BOUNDS ON DIFFERENTIALLY PRIVATE LEARNING VIA COMMUNICATION COMPLEXITY

机译:通过通信复杂性在不同的私人学习上绑定示例复杂性

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

In this work we analyze the sample complexity of classification by differentially private algorithms. Differential privacy is a strong and well-studied notion of privacy introduced by Dwork et al. [Lecture Notes in Comput. Sci. 3876, Springer, New York, 2006, pp. 265-284] that ensures that the output of an algorithm leaks little information about the data point provided by any of the participating individuals. Sample complexity of private probably approximately correct (PAC) and agnostic learning was studied in a number of prior works starting with Kasiviswanathan et al. [SIAM J. Comput., 40 (2011), pp. 793-826]. However, a number of basic questions remain open [A. Beimel, S. P. Kasiviswanathan, and K. Nissim, Lecture Notes in Comput. Sci. 5978, Springer, New York, 2006, pp. 437-454; K. Chaudhuri and D. Hsu, Proceedings of Conference in Learning Theory, 2011, pp. 155-186; A. Beimel, K. Nissim, and U. Stemmer, Proceedings of the 4th Conference on Innovations in Theoretical Computer Science, 2013, pp. 97-110; A. Beimel, K. Nissim, and U. Stemmer, Proceedings of APPROX+RANDOM, 2013, pp. 363-378], most notably whether learning with privacy requires more samples than learning without privacy. We show that the sample complexity of learning with (pure) differential privacy can be arbitrarily higher than the sample complexity of learning without the privacy constraint or the sample complexity of learning with approximate differential privacy. Our second contribution and the main tool is an equivalence between the sample complexity of (pure) differentially private learning of a concept class C (SCDP(C)) and the randomized one-way communication complexity of the evaluation problem for concepts from C. Using this equivalence we prove the following bounds: (a) SCDP(C) = O(LDim(C)), where LDim(C) is the Littlestone's dimension characterizing the number of mistakes in the online-mistakebound learning model [N. Littlestone, Machine Learning, 2 (1987), pp. 285-318]. Known bounds on LDim(C) then imply that SCDP(C) can be much higher than the Vapnik-Chervonenkis dimension of C. (b) For any t, there exists a class C such that LDim(C) = 2 but SCDP(C) >= t. (c) For any t, there exists a class C such that the sample complexity of (pure) a-differentially private PAC learning is Omega(t/alpha) but the sample complexity of the approximate (alpha, beta)-differentially private PAC learning is O(log(1/beta)/alpha). This resolves an open problem from Beimel, Nissim, and Stemmer [Proceedings of APPROX+RANDOM, 2013, pp. 363-378].
机译:在这项工作中,我们通过差分私有算法来分析分类的样本复杂性。差异性隐私是Dwork等人引入的一种强有力且经过充分研究的隐私概念。 [计算笔记。科学[3876,Springer,New York,2006,pp.265-284],它确保算法的输出几乎不泄漏任何参与人员提供的有关数据点的信息。从Kasiviswanathan等人开始的许多先前研究中,都对私人大概近似正确(PAC)和不可知论学习的样本复杂性进行了研究。 [SIAM J.Comput。,40(2011),第793-826页]。但是,仍有许多基本问题[A. Beimel,S.P。Kasiviswanathan和K.Nissim,《计算机讲义》。科学5978,Springer,纽约,2006年,第437-454页。 K. Chaudhuri和D. Hsu,《学习理论会议论文集》,2011年,第155-186页; A. Beimel,K。Nissim和U. Stemmer,第四届理论计算机科学创新会议论文集,2013年,第97-110页; A. Beimel,K。Nissim和U. Stemmer,《 APPROX + RANDOM》,2013年,第363-378页],最值得注意的是,与没有隐私的学习相比,有隐私的学习是否需要更多的样本。我们表明,具有(纯)差异隐私的学习样本复杂度可以比没有隐私约束的学习样本复杂度或具有近似差异隐私的学习样本复杂度任意高。我们的第二个贡献和主要工具是对概念类C的(纯)差分私人学习的样本复杂度(SCDP(C))与对概念C的评估问题的随机单向通信复杂度之间的等价性。这种等效性证明了以下界限:(a)SCDP(C)= O(LDim(C)),其中LDim(C)是Littlestone的维度,用于表征在线错误学习模型中错误的数量[N. Littlestone,《机器学习》,第2卷,1987年,第285-318页。则LDim(C)的已知界限意味着SCDP(C)可以比C的Vapnik-Chervonenkis维数高得多。(b)对于任何t,都存在一个类C,使得LDim(C)= 2但SCDP( C)> = t。 (c)对于任何t,都存在一个C类,使得(纯)a微分专用PAC学习的样本复杂度为Omega(t / alpha),但近似(alpha,beta)微分专用PAC的样本复杂度学习是O(log(1 / beta)/ alpha)。这解决了来自Beimel,Nissim和Stemmer的公开问题[APPROX + RANDOM的议事录,2013年,第363-378页]。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号