首页> 外文会议>AISB Convention >Baselines for Joint-Action Reinforcement Learning of Coordination in Cooperative Multi-Agent Systems
【24h】

Baselines for Joint-Action Reinforcement Learning of Coordination in Cooperative Multi-Agent Systems

机译:联合动作加固学习协调中协调的基线

获取原文

摘要

A common assumption for the study of reinforcement learning of coordination is that agents can observe each other's actions (so-called joint-action learning). We present in this paper a number of simple joint-action learning algorithms and show that they perform very well when compared against more complex approaches such as OAL (Wang and Sandholm, 2002), while still maintaining convergence guarantees. Based on the empirical results, we argue that these simple algorithms should be used as baselines for any future research on joint-action learning of coordination.
机译:对协调的加强学习研究的共同假设是,药剂可以观察彼此的行为(所谓的联合行动学习)。我们在本文中展示了许多简单的联合动作学习算法,并表明它们在与OAL(Wang和Sandholm,2002)等更复杂的方法相比时表现得非常好,同时仍保持收敛保证。基于经验结果,我们认为这些简单的算法应用作基于基准的基础,以便在联合行动学习协调的任何研究中使用基线。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号