首页> 外文会议>SIAM International Conference on Data Mining >Learning Markov Network Structure using Few Independence Tests
【24h】

Learning Markov Network Structure using Few Independence Tests

机译:使用少量独立测试学习马尔可夫网络结构

获取原文
获取外文期刊封面目录资料

摘要

In this paper we present the Dynamic Grow-Shrink Inference-based Markov network learning algorithm (abbreviated DGSIMN), which improves on GSIMN, the state-of-the-art algorithm for learning the structure of the Markov network of a domain from independence tests on data. DGSIMN, like other independence-based algorithms, works by conducting a series of statistical conditional independence tests toward the goal of restricting the number of possible structures to one, thus inferring that structure as the only possibly correct one. During this process, DGSIMN, like the GSIMN algorithm, uses the axioms that govern the probabilistic independence relation to avoid unnecessary tests i.e., tests that can be inferred from the results of known ones. This results in both efficiency and reliability advantages over the simple application of statistical tests. However, one weakness of GSIMN is its rigid and heuristic ordering of the execution of tests, which results in potentially inefficient execution. DGSIMN instead uses a principled strategy, dynamically selecting the locally optimal test that is expected to increase the state of our knowledge about the structure the most. This is done by calculating the expected number of independence facts that will become known (through inference) after executing a particular test (before it is actually evaluated on data), and by selecting the one that is expected to maximize the number of such inferences, thus avoiding their potentially expensive evaluation on data. As we demonstrate in our experiments, this results in an overall decrease in the computational requirements of the algorithm, sometimes dramatically, due to the decreased the number of tests required to be evaluated on data. Experiments show that DGSIMN yields savings of up to 88% on both sampled and benchmark data while achieving similar or better accuracy in most cases.
机译:在本文中,我们介绍了基于动态生长的缩小推断的马尔可夫网络学习算法(缩写DGSIMN),其改进了GSIMN,用于学习域的Markov网络结构的最先进的算法从独立测试关于数据。 DGSIMN与其他基于独立的算法相同,通过对将可能结构的数量限制为一个的目标进行一系列统计条件独立性测试,从而推断该结构作为唯一可能校正的结构。在此过程中,与GSIMN算法一样,DGSIMN使用管理概率独立关系的公理以避免不必要的测试,即,可以从已知结果的结果推断出来的测试。这导致效率和可靠性优势,在简单地应用统计测试方面。然而,GSIMN的一个弱点是其执行测试的刚性和启发式的排序,这导致潜在的执行效率低。 DGSIMN反而使用原则性的策略,动态选择了当地最佳测试,这些测试预计将增加对结构的知识状态最多。这是通过计算在执行特定测试之后(通过推理的推理)(在实际评估数据)之前已知(通过推理)的预期独立事实来完成的,并且通过选择预期最大化此类推广的数量的那个,因此,避免了对数据的潜在昂贵的评估。正如我们在我们的实验中展示,这导致算法的计算要求总体上减少,有时会导致逐渐变化,因为减少了对数据进行评估所需的测试数量。实验表明,在大多数情况下,DGSIMN在采样和基准数据上产生高达88%的节省高达88%,同时在大多数情况下实现相似或更好的准确性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号