首页> 外文会议>International Conference on Autonomous Agents and Multiagent Systems >Efficient Convention Emergence through Decoupled Reinforcement Social Learning with Teacher-Student Mechanism
【24h】

Efficient Convention Emergence through Decoupled Reinforcement Social Learning with Teacher-Student Mechanism

机译:通过与教师 - 学生机制脱钩加强社会学习的有效公约出现

获取原文

摘要

In this paper, we design reinforcement learning based (RL-based) strategies to promote convention emergence in multiagent systems (MASs) with large convention space. We apply our approaches to a language coordination problem in which agents need to coordinate on a dominant lexicon for efficient communication. By modeling each lexicon which maps each concept to a single word as a Markov strategy representation, the original single-state convention learning problem can be transformed into a multi-state multiagent coordination problem. The dynamics of lexicon evolutions during an interaction episode can be modeled as a Markov game, which allows agents to improve the action values of each concept separately and incrementally. Specifically we propose two learning strategies, multiple-Q and multiple-R, and also propose incorporating teacher-student mechanism on top of the learning strategies to accelerate lexicon convergence speed. Extensive experiments verify that our approaches outperform the state-of-the-art approaches in terms of convergence efficiency, convention quality and scalability.
机译:在本文中,我们设计了基于加强学习的(基于RL的)策略,促进了大型会议空间的多元系统(质量)中的公约出现。我们将我们的方法应用于语言协调问题,其中代理商需要协调主导lexicon以进行高效沟通。通过将每个概念映射到单个单词作为Markov策略表示的每个概念的建模,原始单态会议学习问题可以转换为多状态的多态协调问题。交互剧集期间词典演进的动态可以被建模为马尔可夫游戏,这允许代理分别和逐步提高每个概念的动作值。具体而言,我们提出了两个学习策略,多Q和多重r,并提出将师生机制纳入学习策略,以加速词汇融合速度。广泛的实验验证了我们的方法在收敛效率,公约质量和可扩展性方面优于最先进的方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号