首页> 外文会议>American Control Conference >A Simple Learning Rule in Games and Its Convergence to Pure-Strategy Nash Equilibria
【24h】

A Simple Learning Rule in Games and Its Convergence to Pure-Strategy Nash Equilibria

机译:游戏中简单的学习规则及其对纯策略纳什均衡的融合

获取原文

摘要

We propose a simple learning rule in games. The proposed rule only requires that (i) if there exists at least one strictly better reply (SBR), an agent switches its action to each SBR with positive probability or stay with the same action (with positive probability), and (ii) when there is no SBR, the agent either stays with the previous action or switches to another action that yields the same payoff. We first show that some of existing algorithms (or simple modifications) are special cases of our proposed algorithm. Secondly, we demonstrate that this intuitive rule guarantees almost sure convergence to a pure-strategy Nash equilibrium in a large class of games that we call generalized weakly acyclic games. Finally, we show that the probability that the action profile does not converge to a pure-strategy Nash equilibrium decreases geometrically fast in the aforementioned class of games.
机译:我们提出了一个简单的游戏学习统治。所提出的规则只需要(i)如果至少存在一个严格更好的回复(SBR),则代理将其动作切换到每个SBR,每个SBR具有正概率或保持相同的动作(具有正概率),并且(II)没有SBR,代理要么停留在前一个动作或切换到另一个产生相同回报的动作。首先显示现有算法(或简单的修改)是我们所提出的算法的特殊情况。其次,我们证明这种直观的规则几乎肯定会在一大类游戏中达到纯策略的纳什均衡,我们称之为泛滥的无循环游戏。最后,我们表明,在上述比赛中,动作配置文件不会收敛到纯策略的纳什均衡的概率降低了几何上的几何上。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号