首页> 外文会议>IEEE Conference on Decision and Control >A Distributed Algorithm for Solving a Class of Multi-agent Markov Decision Problems
【24h】

A Distributed Algorithm for Solving a Class of Multi-agent Markov Decision Problems

机译:一种求解一类多代理马尔可夫决策问题的分布式算法

获取原文

摘要

This paper considers a class of infinite horizon Markov decision processes (MTOPs) with multiple decision makers, called agents, and a general joint reward structure, but a special decomposable state/action structure such that each individual agent's actions affect the system's state transitions independently from the actions of all other agents. We introduce the concept of "localization," where each agent need only consider a "local" MDP defined on its own state and action spaces. Based on this localization concept, we propose an iterative distributed algorithm that emulates gradient ascent and which converges to a locally optimal solution for the average reward case. The solution is an "autonomous" joint policy such that each agent's action is based on only its local state.
机译:本文考虑了一类无限的地平线马尔可夫决策过程(MTOPS),具有多个决策者,称为代理商和一般的联合奖励结构,但是一个特殊的可分解状态/行动结构,使得每个代理人的行为独立影响系统的国家过渡。所有其他代理商的行为。我们介绍了“本地化”的概念,其中每个代理只需要考虑在其自己的状态和行动空间上定义的“本地”MDP。基于该本地化概念,我们提出了一种迭代分布式算法,其模拟梯度上升,并将其收敛到平均奖励案例的局部最佳解决方案。该解决方案是“自主”联合政策,使每个代理的行动仅基于其当地国家。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号