首页> 外文学位 >Agent interactions in decentralized environments.
【24h】

Agent interactions in decentralized environments.

机译:分散环境中的代理交互。

获取原文
获取原文并翻译 | 示例

摘要

The decentralized Markov decision process (Dec-POMDP) is a powerful formal model for studying multiagent problems where cooperative, coordinated action is optimal, but each agent acts based on local data alone. Unfortunately, it is known that Dec-POMDPs are fundamentally intractable: they are NEXP-complete in the worst case, and have been empirically observed to be beyond feasible optimal solution.;To get around these obstacles, researchers have focused on special classes of the general Dec-POMDP problem, restricting the degree to which agent actions can interact with one another. In some cases, it has been proven that these sorts of structured forms of interaction can in fact reduce worst-case complexity. Where formal proofs have been lacking, empirical observations suggest that this may also be true for other cases, although less is known precisely.;This thesis unifies a range of this existing work, extending analysis to establish novel complexity results for some popular restricted-interaction models. We also establish some new results concerning cases for which reduced complexity has been proven, showing correspondences between basic structural features and the potential for dimensionality reduction when employing mathematical programming techniques.;As our new complexity results establish that worst-case intractability is more widespread than previously known, we look to new ways of analyzing the potential average-case difficulty of Dec-POMDP instances. As this would be extremely difficult using the tools of traditional complexity theory, we take a more empirical approach. In so doing, we identify new analytical measures that apply to all Dec-POMDPs, whatever their structure. These measures allow us to identify problems that are potentially easier to solve on average, and validate this claim empirically. As we show, the performance of well-known optimal dynamic programming methods correlates with our new measure of difficulty. Finally, we explore the approximate case, showing that our measure works well as a predictor of difficulty there, too, and provides a means of setting algorithm parameters to achieve far more efficient performance.
机译:分散的马尔可夫决策过程(Dec-POMDP)是一个功能强大的正式模型,用于研究多主体问题,其中合作,协调的行动是最佳的,但每个主体仅基于本地数据进行行动。不幸的是,众所周知Dec-POMDP从根本上是难于解决的:在最坏的情况下它们是NEXP完全的,并且从经验上已经观察到它们超出了可行的最佳解决方案。为了解决这些障碍,研究人员将注意力集中在特殊的Dec-POMDP的一般性问题,它限制了代理程序行为可以相互交互的程度。在某些情况下,已经证明,这种类型的结构化交互形式实际上可以减少最坏情况下的复杂性。在缺乏正式证据的情况下,经验观察表明,尽管对其他情况的确切了解很少,但对其他情况也可能是正确的。本论文统一了一系列现有工作,扩展了分析范围,以建立一些流行的受限相互作用的新颖复杂性结果楷模。我们还建立了一些有关已证明降低了复杂性的案例的新结果,显示了采用数学编程技术时基本结构特征与降维潜力之间的对应关系。由于我们的新复杂性结果表明,最坏情况下的可处理性比以前已知,我们寻求分析Dec-POMDP实例潜在平均案例难度的新方法。由于使用传统复杂性理论的工具很难做到这一点,因此我们采用了更为经验的方法。通过这样做,我们确定了适用于所有Dec-POMDP的新分析方法,无论其结构如何。这些措施使我们能够确定平均而言可能更容易解决的问题,并凭经验验证这一说法。正如我们所展示的那样,众所周知的最佳动态规划方法的性能与我们的新难度度量相关。最后,我们探索了近似情况,表明我们的度量也可以很好地预测困难,并提供了一种设置算法参数的方法,以实现更高效率的性能。

著录项

  • 作者

    Allen, Martin William.;

  • 作者单位

    University of Massachusetts Amherst.;

  • 授予单位 University of Massachusetts Amherst.;
  • 学科 Artificial Intelligence.;Computer Science.
  • 学位 Ph.D.
  • 年度 2009
  • 页码 214 p.
  • 总页数 214
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类 人工智能理论;自动化技术、计算机技术;
  • 关键词

  • 入库时间 2022-08-17 11:38:09

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号