首页> 外文会议>IEEE Annual Symposium on Foundations of Computer Science >Efficiently Learning Mixtures of Mallows Models
【24h】

Efficiently Learning Mixtures of Mallows Models

机译:有效学习锦葵模型的混合物

获取原文

摘要

Mixtures of Mallows models are a popular generative model for ranking data coming from a heterogeneous population. They have a variety of applications including social choice, recommendation systems and natural language processing. Here we give the first polynomial time algorithm for provably learning the parameters of a mixture of Mallows models with any constant number of components. Prior to our work, only the two component case had been settled. Our analysis revolves around a determinantal identity of Zagier which was proven in the context of mathematical physics, which we use to show polynomial identifiability and ultimately to construct test functions to peel off one component at a time. To complement our upper bounds, we show information-theoretic lower bounds on the sample complexity as well as lower bounds against restricted families of algorithms that make only local queries. Together, these results demonstrate various impediments to improving the dependence on the number of components. They also motivate the study of learning mixtures of Mallows models from the perspective of beyond worst-case analysis. In this direction, we show that when the scaling parameters of the Mallows models have separation, there are much faster learning algorithms.
机译:Mallows模型的混合物是一种流行的生成模型,用于对来自异类种群的数据进行排名。它们具有多种应用程序,包括社交选择,推荐系统和自然语言处理。在这里,我们给出了第一个多项式时间算法,以可证明地学习具有任意常数组件的Mallows模型混合参数。在我们进行工作之前,仅解决了两个组成部分的案例。我们的分析围绕着Zagier的行列式确定性,该确定性在数学物理学的背景下得到了证明,我们用它来显示多项式的可识别性,并最终构建一次剥离一个分量的测试函数。为了补充我们的上限,我们显示了样本复杂度上的信息理论下限,以及仅进行局部查询的受限算法族的下限。总之,这些结果证明了改善对组件数量的依赖性的各种障碍。他们还从最坏情况分析的角度出发,促进了对Mallows模型混合学习的研究。在这个方向上,我们表明,当Mallows模型的缩放参数分离时,将有更快的学习算法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号