首页> 外文会议>International Conference on Autonomous Agents and Multiagent Systems >Just Add Pepper: Extending Learning Algorithms for Repeated Matrix Games to Repeated Markov Games
【24h】

Just Add Pepper: Extending Learning Algorithms for Repeated Matrix Games to Repeated Markov Games

机译:只需添加Pepper:将重复矩阵游戏扩展到重复马尔可夫游戏的学习算法

获取原文

摘要

Learning in multi-agent settings has recently garnered much interest, the result of which has been the development of somewhat effective multi-agent learning (MAL) algorithms for repeated normal-form games. However, general-purpose MAL algorithms for richer environments, such as general sum repeated stochastic (Markov) games (RSGs), are less advanced. Indeed, previously created MAL algorithms for RSGs are typically successful only when the behavior of as sociates meets specific game theoretic assumptions and when the game is of a particular class (such as zero-sum games). In this paper, we present a new algorithm, called Pepper, that can be used to extend MAL algorithms designed for repeated normal-form games to RSGs. We demonstrate that Pepper creates a family of new algorithms, each of whose asymptotic performance in RSGs is reminiscent of its asymptotic performance in related repeated normal-form games. We also show that some algorithms formed with Pepper outperform existing algorithms in an interesting RSG.
机译:在多代理设置中学习最近获得了很多兴趣,这是一直是开发有点有效的多代理学习(MAL)算法,用于重复的正常形式游戏。然而,用于更丰富的环境的通用MAL算法,例如普通和重复的随机(马尔可夫)游戏(RSG),不太高级。实际上,才为RSGS创建的MAL算法通常仅成功,只有当同社会符合特定的游戏理论假设以及游戏的特定类别时(例如零和游戏)时,才会成功。在本文中,我们提出了一种新的算法,称为Pepper,可用于扩展为重复正常格式游戏设计的MAL算法。我们展示了Pepper创造了一系列新算法,每个算法在RSG中的渐近性表现都让人想起其在相关的反复正常形式游戏中的渐近性能。我们还表明,一些用辣椒形成的算法优于现有的现有算法,以有趣的RSG。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号