首页> 外文会议>International Conference on Autonomous Agents and Multiagent Systems >Training Cooperative Agents for Multi-Agent Reinforcement Learning: Extended Abstract
【24h】

Training Cooperative Agents for Multi-Agent Reinforcement Learning: Extended Abstract

机译:培训合作社多智能经纪增强学习:扩展摘要

获取原文

摘要

Deep Learning and back-propagation has been successfully used to perform centralized training with communication protocols among multiple agents in a cooperative environment. In this paper we present techniques for centralized training of Multi-Agent (Deep) Reinforcement Learning (MARL) using the model-free Deep Q-Network as the baseline model and message sharing between agents. We present a novel, scalable, centralized MARL training technique, which separates the message learning module from the policy module. The separation of these modules helps in faster convergence in complex domains like autonomous driving simulators. A second contribution uses the centrally trained model to bootstrap training of distributed, independent, cooperative agent policies for execution and thus addresses the challenges of noise and communication bottlenecks in real-time communication channels. This paper theoretically and empirically compares our centralized training algorithms to current research in the field of MARL. We also present and release a new OpenAI-Gym environment which can be used for multi-agent research as it simulates multiple autonomous cars driving cooperatively on a highway.
机译:深度学习和背部传播已成功用于在合作环境中使用多个代理之间的通信协议进行集中培训。在本文中,我们使用无模型的深Q-Network作为基线模型和代理之间的消息共享的基线模型和消息共享的集中训练技术来实现多助手(深)加强学习(MARL)的技术。我们提出了一种新颖,可扩展,集中的MARL训练技术,将消息学习模块与策略模块分开。这些模块的分离有助于更快地在自动驾驶模拟器等复杂域中收敛。第二贡献利用集中培训的模型来启动分布式,独立的合作社政策的启动培训,从而解决了实时通信渠道中噪声和通信瓶颈的挑战。本文理论上并经验与我们的集中式训练算法进行了对Marl领域的当前研究。我们还介绍并释放了一个新的Openai-Mave环境,可用于多智能经纪研究,因为它模拟了在高速公路上协同驾驶的多个自主汽车。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号