首页> 外文会议>2011 19th Mediterranean Conference on Control Automation (MED) >Decentralized learning in multiple pursuer-evader Markov games
【24h】

Decentralized learning in multiple pursuer-evader Markov games

机译:在多个追逐者马氏游戏中的去中心化学习

获取原文
获取外文期刊封面目录资料

摘要

We represent the multiple pursuers and evaders game as a Markov game and each player as a decentralized unit that has to work independently in order to complete a task. Most proposed solutions for this distributed multiagent decision problem require some sort of central coordination. In this paper, we intend to model each player as a learning automata (LA) and let them evolve and adapt in order to solve the difficult problem they have at hand. We are also going to show that using the proposed learning process, the players' policies will converge to an equilibrium point. Simulations of such scenarios with multiple pursuers and evaders are presented in order to show the feasibility of the approach.
机译:我们将多个追随者和逃避者游戏表示为马尔可夫游戏,而每个玩家则作为分散的单位来代表,这些单位必须独立工作才能完成任务。针对此分布式多主体决策问题提出的大多数解决方案都需要某种中央协调。在本文中,我们打算将每个参与者建模为学习自动机(LA),让他们发展和适应,以解决他们面临的难题。我们还将证明,通过提议的学习过程,参与者的政策将收敛到一个平衡点。为了证明这种方法的可行性,提出了具有多个追踪者和逃避者的这种情况的仿真。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号