首页> 外文会议>IEEE Annual Symposium on Foundations of Computer Science >Efficiently Learning Mixtures of Mallows Models
【24h】

Efficiently Learning Mixtures of Mallows Models

机译:有效地学习Mallows模型的混合物

获取原文

摘要

Mixtures of Mallows models are a popular generative model for ranking data coming from a heterogeneous population. They have a variety of applications including social choice, recommendation systems and natural language processing. Here we give the first polynomial time algorithm for provably learning the parameters of a mixture of Mallows models with any constant number of components. Prior to our work, only the two component case had been settled. Our analysis revolves around a determinantal identity of Zagier which was proven in the context of mathematical physics, which we use to show polynomial identifiability and ultimately to construct test functions to peel off one component at a time. To complement our upper bounds, we show information-theoretic lower bounds on the sample complexity as well as lower bounds against restricted families of algorithms that make only local queries. Together, these results demonstrate various impediments to improving the dependence on the number of components. They also motivate the study of learning mixtures of Mallows models from the perspective of beyond worst-case analysis. In this direction, we show that when the scaling parameters of the Mallows models have separation, there are much faster learning algorithms.
机译:Mallows模型的混合物是一种流行的生成模型,用于排名来自异构人群的数据。它们有各种应用,包括社交选择,推荐系统和自然语言处理。在这里,我们给出了第一种多项式时间算法,用于以任何恒定数量的组件来证明用于可证明的Mallows模型混合参数的参数。在我们工作之前,只有两个组成案件已经解决。我们的分析周围围绕了Zagier的决定性身份,这在数学物理学的背景下被证明,我们用于显示多项式可识别性,最终构建测试功能一次剥离一个组件。为了补充我们的上限,我们向样本复杂度的信息 - 理论下限以及仅对当地查询的限制算法的受限制家庭的下限。这些结果一起表明了改善对组件数量的依赖性的各种障碍。他们还可以从超出最坏情况分析的角度来激励抹菜模型的学习混合物。在这个方向上,我们表明,当Mallows模型的缩放参数具有分离时,有更快的学习算法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号