首页> 外文会议>20th European conference on artificial intelligence >A Reinforcement-Learning Algorithm for Sampling Design in Markov Random Fields
【24h】

A Reinforcement-Learning Algorithm for Sampling Design in Markov Random Fields

机译:马氏随机场中抽样设计的强化学习算法

获取原文
获取原文并翻译 | 示例

摘要

Optimal sampling in spatial random fields is a complex problem, which mobilizes several research fields in spatial statistics and artificial intelligence. In this paper we consider the case where observations are discrete-valued and modelled by a Markov Random Field. Then we encode the sampling problem into the Markov Decision Process (MDP) framework. After exploring existing heuristic solutions as well as classical algorithms from the field of Reinforcement Learning (RL), we design an original algorithm, LSDP (Least Square Dynamic Programming), which uses simulated trajectories to solve approximately any finite-horizon MDP problem. Based on an empirical study of the behaviour of these different approaches on binary models, we derive the following conclusions: i) a naive heuristic, consisting in sampling sites where marginals are the most uncertain, is already an efficient sampling approach; ii) LSDP outperforms all the classical RL approaches we have tested; iii) LSDP outperforms the heuristic in cases when reconstruction errors have a high cost, or sampling actions are constrained. In addition, LSDP readily handles action costs in the optimisation problem, as well as cases when some sites of the MRF can not be observed.
机译:空间随机场中的最优采样是一个复杂的问题,它动员了空间统计和人工智能领域的几个研究领域。在本文中,我们考虑了观测值是离散值并由马尔可夫随机场建模的情况。然后,我们将采样问题编码到马尔可夫决策过程(MDP)框架中。在探究了现有的启发式解决方案以及强化学习(RL)领域的经典算法之后,我们设计了一种原始算法LSDP(最小二乘动态规划),该算法使用模拟轨迹来解决大约任何有限水平MDP问题。基于对这些不同方法在二元模型上的行为的经验研究,我们得出以下结论:i)朴素的启发式方法(由边际最不确定的采样点组成)已经是一种有效的采样方法; ii)LSDP优于我们测试过的所有经典RL方法; iii)在重建误差代价高昂或采样动作受到限制的情况下,LSDP优于启发式算法。此外,LSDP可以轻松处理优化问题中的动作成本,以及无法观察到MRF某些位置的情况。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号