...
首页> 外文期刊>Journal of machine learning research >Learning Hidden Variable Networks: The Information Bottleneck Approach
【24h】

Learning Hidden Variable Networks: The Information Bottleneck Approach

机译:学习隐藏变量网络:信息瓶颈方法

获取原文

摘要

A central challenge in learning probabilistic graphical models is dealing with domains that involve hidden variables. The common approach for learning model parameters in such domains is the expectation maximization (EM) algorithm. This algorithm, however, can easily get trapped in sub-optimal local maxima. Learning the model structure is even more challenging. The structural EM algorithm can adapt the structure in the presence of hidden variables, but usually performs poorly without prior knowledge about the cardinality and location of the hidden variables. In this work, we present a general approach for learning Bayesian networks with hidden variables that overcomes these problems. The approach builds on the information bottleneck framework of Tishby et al. (1999). We start by proving formal correspondence between the information bottleneck objective and the standard parametric EM functional. We then use this correspondence to construct a learning algorithm that combines an information-theoretic smoothing term with a continuation procedure. Intuitively, the algorithm bypasses local maxima and achieves superior solutions by following a continuous path from a solution of, an easy and smooth, target function, to a solution of the desired likelihood function. As we show, our algorithmic framework allows learning of the parameters as well as the structure of a network. In addition, it also allows us to introduce new hidden variables during model selection and learn their cardinality. We demonstrate the performance of our procedure on several challenging real-life data sets. color="gray">
机译:学习概率图形模型的主要挑战是处理涉及隐藏变量的域。在这种领域中学习模型参数的常用方法是期望最大化算法。但是,该算法很容易陷入次优的局部最大值。学习模型结构更具挑战性。 结构EM 算法可以在存在隐藏变量的情况下适应结构,但是在没有有关隐藏变量的基数和位置的先验知识的情况下,其性能通常很差。在这项工作中,我们提出了一种克服这些问题的,具有隐藏变量的贝叶斯网络学习方法。该方法建立在Tishby等人的信息瓶颈框架上。 (1999)。我们首先证明信息瓶颈目标和标准参数EM功能之间的形式对应。然后,我们使用这种对应关系来构造学习算法,该算法将信息理论的平滑项与延续过程相结合。直观地讲,该算法绕过了局部极大值,并遵循从简单而平稳的目标函数的解到期望似然函数的解的连续路径,从而获得了出色的解。正如我们所展示的,我们的算法框架允许学习参数以及网络的结构。此外,它还允许我们在模型选择期间引入新的隐藏变量并了解其基数。我们在一些具有挑战性的现实数据集上演示了该过程的性能。 color =“ gray”>

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号