...
首页> 外文期刊>The Journal of Artificial Intelligence Research >On Polynomial Sized MDP Succinct Policies
【24h】

On Polynomial Sized MDP Succinct Policies

机译:关于多项式大小的MDP简洁策略

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Policies of Markov Decision Processes (MDPs) determine the next action to execute from the current state and, possibly, the history (the past states). When the number of states is large, succinct representations are often used to compactly represent both the MDPs and the policies in a reduced amount of space. In this paper, some problems related to the size of succinctly represented policies are analyzed. Namely, it is shown that some MDPs have policies that can only be represented in space super-polynomial in the size of the MDP, unless the polynomial hierarchy collapses. This fact motivates the study of the problem of deciding whether a given MDP has a policy of a given size and reward. Since some algorithms for MDPs work by finding a succinct representation of the value function, the problem of deciding the existence of a succinct representation of a value function of a given size and reward is also considered.
机译:马尔可夫决策过程(MDP)的策略根据当前状态以及可能的历史记录(过去的状态)来确定要执行的下一个动作。当状态数很大时,通常使用简洁的表示形式在较小的空间中紧凑地表示MDP和策略。本文分析了与简洁代表的政策规模有关的一些问题。即,示出了某些MDP具有只能以MDP的大小在空间超多项式中表示的策略,除非多项式层次结构崩溃。这一事实激发了对确定给定MDP是否具有给定规模和奖励策略的问题的研究。由于用于MDP的某些算法通过找到价值函数的简洁表示来工作,因此还考虑了确定给定大小和奖励的价值函数的简洁表示存在的问题。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号