首页> 外文期刊>Journal of machine learning research >A Bayesian Approach for Learning and Planning in Partially Observable Markov Decision Processes
【24h】

A Bayesian Approach for Learning and Planning in Partially Observable Markov Decision Processes

机译:在局部可观的马尔可夫决策过程中进行学习和计划的贝叶斯方法

获取原文
           

摘要

Bayesian learning methods have recently been shown to provide an elegant solution to the exploration-exploitation trade-off in reinforcement learning. However most investigations of Bayesian reinforcement learning to date focus on the standard Markov Decision Processes (MDPs). The primary focus of this paper is to extend these ideas to the case of partially observable domains, by introducing the Bayes-Adaptive Partially Observable Markov Decision Processes. This new framework can be used to simultaneously (1) learn a model of the POMDP domain through interaction with the environment, (2) track the state of the system under partial observability, and (3) plan (near-)optimal sequences of actions. An important contribution of this paper is to provide theoretical results showing how the model can be finitely approximated while preserving good learning performance. We present approximate algorithms for belief tracking and planning in this model, as well as empirical results that illustrate how the model estimate and agent's return improve as a function of experience. color="gray">
机译:最近显示出贝叶斯学习方法为强化学习中的探索与开发权衡提供了一种优雅的解决方案。但是,迄今为止,对贝叶斯强化学习的大多数研究都集中在标准的马尔可夫决策过程(MDP)上。本文的主要重点是通过介绍贝叶斯自适应部分可观察的马尔可夫决策过程,将这些思想扩展到部分可观察的域。此新框架可用于同时(1)通过与环境交互来学习POMDP域的模型;(2)在部分可观察性的情况下跟踪系统状态;以及(3)计划(近乎)最佳操作序列。本文的重要贡献是提供理论结果,说明如何在保持良好学习性能的同时对模型进行有限近似。我们提供此模型中用于信念跟踪和计划的近似算法,以及经验结果,这些结果说明了模型估计和代理回报如何根据经验而提高。 color =“ gray”>

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号