【24h】

Training Factor Graphs with Reinforcement Learning for Efficient MAP Inference

机译:通过强化学习训练因子图以实现有效的MAP推理

获取原文

摘要

Large, relational factor graphs with structure defined by first-order logic or other languages give rise to notoriously difficult inference problems. Because unrolling the structure necessary to represent distributions over all hypotheses has exponential blow-up, solutions are often derived from MCMC. However, because of limitations in the design and parameterization of the jump function, these sampling-based methods suffer from local minima-the system must transition through lower-scoring configurations before arriving at a better MAP solution. This paper presents a new method of explicitly selecting fruitful downward jumps by leveraging reinforcement learning (RL). Rather than setting parameters to maximize the likelihood of the training data, parameters of the factor graph are treated as a log-linear function approximator and learned with methods of temporal difference (TD); MAP inference is performed by executing the resulting policy on held out test data. Our method allows efficient gradient updates since only factors in the neighborhood of variables affected by an action need to be computed-we bypass the need to compute marginals entirely. Our method yields dramatic empirical success, producing new state-of-the-art results on a complex joint model of ontology alignment, with a 48% reduction in error over state-of-the-art in that domain.
机译:具有一阶逻辑或其他语言定义的结构的大型关系因子图引起了非常困难的推理问题。因为展开表示所有假设的分布所必需的结构具有指数爆炸性,所以解决方案通常来自MCMC。但是,由于跳转函数的设计和参数化方面的限制,这些基于采样的方法存在局部最小值的问题,系统必须在获得更好的MAP解决方案之前通过评分较低的配置进行转换。本文提出了一种通过利用强化学习(RL)明确选择富有成效的向下跳跃的新方法。不是将参数设置为最大化训练数据的可能性,而是将因子图的参数视为对数线性函数逼近器,并使用时差(TD)方法进行学习; MAP推断是通过对保留的测试数据执行结果策略来执行的。我们的方法允许有效的梯度更新,因为仅需要计算受动作影响的变量附近的因子,而无需完全计算边际。我们的方法取得了巨大的经验成功,在复杂的本体对齐联合模型上产生了最新的技术成果,与该领域的最新技术相比,误差降低了48%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号