...
首页> 外文期刊>Journal of Statistical Physics >Learning Probabilities From Random Observables in High Dimensions: The Maximum Entropy Distribution and Others
【24h】

Learning Probabilities From Random Observables in High Dimensions: The Maximum Entropy Distribution and Others

机译:从高维度的随机可观测值中学习概率:最大熵分布及其他

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

We consider the problem of learning a target probability distribution over a set of N binary variables from the knowledge of the expectation values (with this target distribution) of M observables, drawn uniformly at random. The space of all probability distributions compatible with these M expectation values within some fixed accuracy, called version space, is studied. We introduce a biased measure over the version space, which gives a boost increasing exponentially with the entropy of the distributions and with an arbitrary inverse 'temperature' . The choice of allows us to interpolate smoothly between the unbiased measure over all distributions in the version space () and the pointwise measure concentrated at the maximum entropy distribution (). Using the replica method we compute the volume of the version space and other quantities of interest, such as the distance R between the target distribution and the center-of-mass distribution over the version space, as functions of and for large N. Phase transitions at critical values of are found, corresponding to qualitative improvements in the learning of the target distribution and to the decrease of the distance R. However, for fixed , the distance R does not vary with , which means that the maximum entropy distribution is not closer to the target distribution than any other distribution compatible with the observable values. Our results are confirmed by Monte Carlo sampling of the version space for small system sizes ().
机译:我们考虑从对M个可观察变量的期望值(具有该目标分布)的知识中随机分布的N个二元变量集学习目标概率分布的问题。研究了在一定的固定精度内与这些M个期望值兼容的所有概率分布的空间,称为版本空间。我们在版本空间上引入了一个偏差量度,该量度随着分布的熵以及任意的逆“温度”而呈指数增长。选择允许我们在版本空间中所有分布的无偏度量和集中于最大熵分布的逐点度量之间进行平滑插值。使用复制方法,我们计算版本空间的体积和其他感兴趣的数量,例如目标空间与版本空间上的质心分布之间的距离R,以及作为大N的函数。相变找到临界值时,对应于目标分布学习中的质性改进和距离R的减小。但是,对于固定值,距离R不会随改变,这意味着最大熵分布不会更接近目标分布比与可观察值兼容的任何其他分布。对于小型系统尺寸(),蒙特卡洛(Monte Carlo)对版本空间进行了抽样验证了我们的结果。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号