首页> 外文会议>IEEE/WIC/ACM International Conference on Intelligent Agent Technology >Decomposing large-scale POMDP via belief state analysis
【24h】

Decomposing large-scale POMDP via belief state analysis

机译:通过信仰状态分析分解大规模的POMDP

获取原文

摘要

Partially observable Markov decision process (POMDP) is commonly used to model a stochastic environment with unobservable states for supporting optimal decision making. Computing the optimal policy for a large-scale POMDP is known to be intractable. Belief compression, being an approximate solution, has recently been proposed to reduce the dimension of POMDP's belief state space and shown to be effective in improving the problem tractability. In this paper, with the conjecture that temporally close belief states could be characterized by a lower intrinsic dimension, we propose a spatio-temporal brief clustering that considers both the belief states' spatial (in the belief space) and temporal similarities, as well as incorporate it into the belief compression algorithm. The proposed clustering results in belief state clusters as sub-POMDPs of much lower dimension so as to be distributed to a set of distributed agents for collaborative problem solving. The proposed method has been tested using a synthesized navigation problem (Hallway2) and empirically shown to be able to result in policies of superior long-term rewards when compared with those based on solely belief compression. Some future research directions for extending this belief state analysis approach are also included.
机译:部分观察到的马尔可夫决策过程(POMDP)通常用于模拟随意的状态,以支持最佳决策。已知计算大规模POMDP的最佳策略是难以相解的。最近提出了近似解决方案的信念压缩,以减少POMDP信仰状态空间的维度,并显示有效地改善问题途径。在本文中,随着猜想的情况下,时间近乎信仰状态可以通过较低的内在维度来表征,提出了一种审视信仰状态的时空短暂的聚类,以既有信仰国家的空间(在信仰空间)和时间相似之处,以及将其与信仰压缩算法合并。所提出的聚类导致信仰状态集群作为低于尺寸的子POMDP,以便分布到一组分布式代理以进行协作问题。已经使用合成导航问题(Hallway2)测试了所提出的方法,并经验证明能够在与基于完全信仰压缩的那些相比时能够导致优越的长期奖励的政策。还包括一些未来的延长这种信仰状态分析方法的研究方向。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号