首页> 外文会议>Conference on uncertainty in artificial intelligence >Scaling Up Decentralized MDPs Through Heuristic Search
【24h】

Scaling Up Decentralized MDPs Through Heuristic Search

机译:通过启发式搜索扩大去中心化MDP

获取原文

摘要

Decentralized partially observable Markov decision processes (Dec-POMDPs) are rich models for cooperative decision-making under uncertainty, but are often intractable to solve optimally (NEXP-complete). The transition and observation independent Dec-MDP is a general subclass that has been shown to have complexity in NP, but optimal algorithms for this subclass are still inefficient in practice. In this paper, we first provide an updated proof that an optimal policy does not depend on the histories of the agents, but only the local observations. We then present a new algorithm based on heuristic search that is able to expand search nodes by using constraint optimization. We show experimental results comparing our approach with the state-of-the-art Dec-MDP and Dec-POMDP solvers. These results show a reduction in computation time and an increase in scalability by multiple orders of magnitude in a number of benchmarks.
机译:分散的,部分可观察的马尔可夫决策过程(Dec-POMDPs)是不确定条件下合作决策的丰富模型,但通常很难解决(NEXP-complete)。独立于过渡和观测的Dec-MDP是一个通用子类,已被证明在NP中具有复杂性,但是针对该子类的最佳算法在实践中仍然效率低下。在本文中,我们首先提供更新的证据,即最优政策不取决于代理商的历史,而仅取决于本地观察。然后,我们提出一种基于启发式搜索的新算法,该算法能够通过使用约束优化来扩展搜索节点。我们展示了将我们的方法与最新的Dec-MDP和Dec-POMDP求解器进行比较的实验结果。这些结果表明,在许多基准测试中,计算时间减少了,可扩展性提高了多个数量级。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号