首页> 外文会议>International conference on scalable uncertainty management >Multivariate Cluster-Based Discretization for Bayesian Network Structure Learning
【24h】

Multivariate Cluster-Based Discretization for Bayesian Network Structure Learning

机译:贝叶斯网络结构学习的多元聚类离散化

获取原文

摘要

While there exist many efficient algorithms in the literature for learning Bayesian networks with discrete random variables, learning when some variables are discrete and others are continuous is still an issue. A common way to tackle this problem is to preprocess datasets by first discretizing continuous variables and, then, resorting to classical discrete variable-based learning algorithms. However, such a method is inefficient because the conditional dependences/arcs learnt during the learning phase bring valuable information that cannot be exploited by the discretization algorithm, thereby preventing it to be fully effective In this paper, we advocate to discretize while learning and we propose a new multivariate discretization algorithm that takes into account all the conditional dependences/arcs learnt so far. Unlike popular discretization methods, ours does not rely on entropy but on clustering using an EM scheme based on a Gaussian mixture model. Experiments show that our method significantly outperforms the state-of-the-art algorithms.
机译:尽管文献中存在许多用于学习具有离散随机变量的贝叶斯网络的有效算法,但是当某些变量是离散变量而其他变量是连续变量时,学习仍然是一个问题。解决此问题的一种常用方法是通过先离散化连续变量然后再求助于经典的基于离散变量的学习算法对数据集进行预处理。但是,这种方法效率低下,因为在学习阶段学习的条件依赖/弧带来了离散化算法无法利用的有价值的信息,从而使其无法完全有效。在本文中,我们主张在学习时进行离散化,并提出一种新的多元离散化算法,该算法考虑了到目前为止学习到的所有条件相关性/弧度。与流行的离散化方法不同,我们的方法不依赖于熵,而是依赖于使用基于高斯混合模型的EM方案进行聚类。实验表明,我们的方法明显优于最新算法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号