...
首页> 外文期刊>IEEE Transactions on Information Theory >Forest Learning From Data and its Universal Coding
【24h】

Forest Learning From Data and its Universal Coding

机译:森林从数据中学习及其通用编码

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

This paper considers structure learning from data with n samples of p variables, assuming that the structure is a forest, using the Chow-Liu algorithm. Specifically, for incomplete data, we construct two model selection algorithms that complete in O(p2) steps: one obtains a forest with the maximum posterior probability given the data and the other obtains a forest that converges to the true one as n increases. We show that the two forests are generally different when some values are missing. In addition, we present estimations for benchmark data sets to demonstrate that both algorithms work in realistic situations. Moreover, we derive the conditional entropy provided that no value is missing, and we evaluate the per-sample expected redundancy for the universal coding of incomplete data in terms of the number of non-missing samples.
机译:本文使用Chow-Liu算法考虑了从具有p个变量的n个样本的数据中进行结构学习,假设结构是森林。具体来说,对于不完整的数据,我们构造了两个模型选择算法,它们以O(p n 2)步骤:一个获得了具有给定数据的最大后验概率的森林,另一个获得了收敛到真实值的森林n增加一个。我们表明,缺少某些值时,这两个森林通常是不同的。此外,我们提出了基准数据集的估计值,以证明这两种算法都可以在现实情况下工作。此外,我们推导条件熵,前提是不遗漏任何值,并根据非缺失样本的数量来评估不完整数据的通用编码的每个样本的预期冗余。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号