首页> 外文会议>IEEE International Conference on Electro Information Technology >Deep Deterministic Policy Gradients with Transfer Learning Framework in StarCraft Micromanagement
【24h】

Deep Deterministic Policy Gradients with Transfer Learning Framework in StarCraft Micromanagement

机译:《星际争霸》微管理中具有转移学习框架的深度确定性策略梯度

获取原文

摘要

This paper proposes an intelligent multi-agent approach in a real-time strategy game, StarCraft, based on the deep deterministic policy gradients (DDPG) techniques. An actor and a critic network are established to estimate the optimal control actions and corresponding value functions, respectively. A special reward function is designed based on the agents' own condition and enemies' information to help agents make intelligent control in the game. Furthermore, in order to accelerate the learning process, the transfer learning techniques are integrated into the training process. Specifically, the agents are trained initially in a simple task to learn the basic concept for the combat, such as detouring moving, avoiding and joining attacking. Then, we transfer this experience to the target task with a complex and difficult scenario. From the experiment, it is shown that our proposed algorithm with transfer learning can achieve better performance.
机译:本文基于深度确定性策略梯度(DDPG)技术,在实时策略游戏StarCraft中提出了一种智能多代理方法。建立演员和评论家网络以分别估计最佳控制动作和相应的价值函数。根据特工的自身状况和敌人的信息设计特殊的奖励功能,以帮助特工在游戏中进行智能控制。此外,为了加快学习过程,将转移学习技术集成到培训过程中。具体而言,首先在一个简单的任务中对特工进行培训,以学习战斗的基本概念,例如绕行moving回,躲避和加入进攻。然后,我们将这种经验转移到复杂而困难的情况下的目标任务中。从实验中可以看出,我们提出的带有转移学习的算法可以取得更好的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号