首页> 美国卫生研究院文献>Entropy >Estimating the Mutual Information between Two Discrete Asymmetric Variables with Limited Samples
【2h】

Estimating the Mutual Information between Two Discrete Asymmetric Variables with Limited Samples

机译:估计具有有限样本的两个离散非对称变量之间的相互信息

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Determining the strength of nonlinear, statistical dependencies between two variables is a crucial matter in many research fields. The established measure for quantifying such relations is the mutual information. However, estimating mutual information from limited samples is a challenging task. Since the mutual information is the difference of two entropies, the existing Bayesian estimators of entropy may be used to estimate information. This procedure, however, is still biased in the severely under-sampled regime. Here, we propose an alternative estimator that is applicable to those cases in which the marginal distribution of one of the two variables—the one with minimal entropy—is well sampled. The other variable, as well as the joint and conditional distributions, can be severely undersampled. We obtain a consistent estimator that presents very low bias, outperforming previous methods even when the sampled data contain few coincidences. As with other Bayesian estimators, our proposal focuses on the strength of the interaction between the two variables, without seeking to model the specific way in which they are related. A distinctive property of our method is that the main data statistics determining the amount of mutual information is the inhomogeneity of the conditional distribution of the low-entropy variable in those states in which the large-entropy variable registers coincidences.
机译:确定两个变量之间的非线性,统计依赖性的强度是许多研究领域的关键问题。用于量化此类关系的既定措施是相互信息。然而,估计有限样本的互信息是一个具有挑战性的任务。由于互信息是两个熵的差异,因此熵的现有贝叶斯估计器可用于估计信息。然而,此程序仍然偏向于受到严重的欠采样的制度。在这里,我们提出了一种替代估计,适用于其中两个变量之一的边缘分布 - 具有最小熵的案例 - 是良好的采样。另一个变量以及关节和条件分布可以严重缺乏采样。我们获得了一致的估计器,即使当采样的数据包含几个巧合时,也表现出非常低的偏差,优于先前的方法。与其他贝叶斯估计人一样,我们的建议侧重于两个变量之间相互作用的强度,而无需寻求建模他们与之相关的具体方式。我们的方法的一个独特性属性是确定互信息的主要数据统计数据是那些大熵变量寄存器重合的那些状态下的低熵变量的不均匀性。

著录项

  • 期刊名称 Entropy
  • 作者单位
  • 年(卷),期 2019(21),6
  • 年度 2019
  • 页码 623
  • 总页数 20
  • 原文格式 PDF
  • 正文语种
  • 中图分类
  • 关键词

    机译:贝叶斯估计;相互信息;偏见;抽样;

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号