首页> 外文会议>EPIA Conference on Artificial Intelligence >Multi-agent Double Deep Q-Networks
【24h】

Multi-agent Double Deep Q-Networks

机译:多代理双深度Q-Networks

获取原文

摘要

There are many open issues and challenges in the multi-agent reward-based learning field. Theoretical convergence guarantees are lost, and the complexity of the action-space is also exponential to the amount of agents calculating their optimal joint-action. Function approximators, such as deep neural networks, have successfully been used in single-agent environments with high dimensional state-spaces. We propose the Multi-agent Double Deep Q-Networks algorithm, an extension of Deep Q-Networks to the multi-agent paradigm. Two common techniques of multi-agent Q-learning are used to formally describe our proposal, and are tested in a Foraging Task and a Pursuit Game. We also demonstrate how they can generalize to similar tasks and to larger teams, due to the strength of deep-learning techniques, and their viability for transfer learning approaches. With only a small fraction of the initial task's training, we adapt to longer tasks, and we accelerate the task completion by increasing the team size, thus empirically demonstrating a solution to the complexity issues of the multi-agent field.
机译:基于Multi-Agent奖励的学习领域存在许多公开问题和挑战。理论收敛保证丢失,并且动作空间的复杂性也是指数的,其代理量计算了最佳关节作用。功能近似器(例如深神经网络)已成功用于具有高维状态空间的单代理环境中。我们提出了多代理双深度Q网络算法,将深度Q-Networks的扩展到多功率范例。多代理Q学习的两种常见技术用于正式描述我们的建议,并在觅食任务和追求游戏中进行测试。由于深度学习技术的强度及其对转移学习方法的可行性,我们还展示了如何概括为类似的任务和更大的团队。只有初始任务培训的一小部分,我们适应更长的任务,我们通过增加团队规模来加速任务完成,从而凭经验证明了对多智能域的复杂性问题的解决方案。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号