首页> 外文会议>AAAI Workshop on Multiagent Learning >Multiagent Q-Learning: Preliminary Study on Dominance between the Nash and Stackelberg Equilibriums
【24h】

Multiagent Q-Learning: Preliminary Study on Dominance between the Nash and Stackelberg Equilibriums

机译:多元型Q学习:纳什和Stackelberg均衡之间的优势初步研究

获取原文
获取外文期刊封面目录资料

摘要

Some game theory approaches to solve multiagent reinforcement learning in self play, i.e. when agents use the same algorithm for choosing action, employ equilibriums, such as the Nash equilibrium, to compute the policies of the agents. These approaches have been applied only on simple examples. In this paper, we present an extended version of Nash Q-Learning using the Stackelberg equilibrium to address a wider range of games than with the Nash Q-Learning. We show that mixing the Nash and Stackelberg equilibriums can lead to better rewards not only in static games but also in stochastic games. Moreover, we apply the algorithm to a real world example, the automated vehicle coordination problem.
机译:一些博弈论解决自我播放中的多层加固学习的方法,即当代理使用相同的算法来选择动作时,采用均衡,例如纳什均衡,计算代理的策略。这些方法仅应用于简单的例子。在本文中,我们使用Stackelberg均衡介绍了NASH Q-Learning的扩展版本,以满足更广泛的游戏范围而不是Nash Q-Learning。我们展示了纳什和Stackelberg均衡的混合可以导致不仅在静态游戏中更好地奖励,也可以在随机游戏中获得更好的奖励。此外,我们将该算法应用于真实的世界示例,自动化的车辆协调问题。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号