【24h】

Multi-agent Double Deep Q-Networks

机译:多主体双深度Q网络

获取原文

摘要

There are many open issues and challenges in the multi-agent reward-based learning field. Theoretical convergence guarantees are lost, and the complexity of the action-space is also exponential to the amount of agents calculating their optimal joint-action. Function approximators, such as deep neural networks, have successfully been used in single-agent environments with high dimensional state-spaces. We propose the Multi-agent Double Deep Q-Networks algorithm, an extension of Deep Q-Networks to the multi-agent paradigm. Two common techniques of multi-agent Q-learning are used to formally describe our proposal, and are tested in a Foraging Task and a Pursuit Game. We also demonstrate how they can generalize to similar tasks and to larger teams, due to the strength of deep-learning techniques, and their viability for transfer learning approaches. With only a small fraction of the initial task's training, we adapt to longer tasks, and we accelerate the task completion by increasing the team size, thus empirically demonstrating a solution to the complexity issues of the multi-agent field.
机译:在基于多主体奖励的学习领域中存在许多未解决的问题和挑战。理论上的收敛保证丢失了,作用空间的复杂性也与计算其最佳联合作用的主体数量成指数关系。函数逼近器(例如深度神经网络)已成功用于具有高维状态空间的单代理环境。我们提出了多智能体双重深层Q网络算法,这是深层Q网络对多智能体范例的扩展。多主体Q学习的两种常用技术被用来正式描述我们的建议,并在觅食任务和追求游戏中进行了测试。我们还演示了由于深度学习技术的强项及其在迁移学习方法中的可行性,它们如何可以推广到相似的任务和更大的团队。仅需训练初始任务的一小部分,我们就可以适应更长的任务,并且通过增加团队规模来加快任务的完成速度,从而从经验上证明了解决多主体领域的复杂性问题的方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号