首页> 外文会议>IEEE International Conference on Electro Information Technology >Deep Deterministic Policy Gradients with Transfer Learning Framework in StarCraft Micromanagement
【24h】

Deep Deterministic Policy Gradients with Transfer Learning Framework in StarCraft Micromanagement

机译:在星际微臂中转移学习框架的深度确定性政策梯度

获取原文

摘要

This paper proposes an intelligent multi-agent approach in a real-time strategy game, StarCraft, based on the deep deterministic policy gradients (DDPG) techniques. An actor and a critic network are established to estimate the optimal control actions and corresponding value functions, respectively. A special reward function is designed based on the agents' own condition and enemies' information to help agents make intelligent control in the game. Furthermore, in order to accelerate the learning process, the transfer learning techniques are integrated into the training process. Specifically, the agents are trained initially in a simple task to learn the basic concept for the combat, such as detouring moving, avoiding and joining attacking. Then, we transfer this experience to the target task with a complex and difficult scenario. From the experiment, it is shown that our proposed algorithm with transfer learning can achieve better performance.
机译:本文基于深度确定性政策梯度(DDPG)技术,提出了一种实时战略游戏中的智能多智能经验方法,Starcraft。建立一个演员和批评网络,以分别估计最佳控制动作和相应的值函数。专用奖励功能是根据代理商的自己的条件和敌人的信息设计,以帮助代理商在游戏中做出智能控制。此外,为了加速学习过程,转移学习技术被集成到训练过程中。具体而言,该代理最初在简单的任务中训练,以学习战斗的基本概念,例如纠缠移动,避免和加入攻击。然后,我们将此体验转移到目标任务,具有复杂和困难的情景。从实验开始,显示我们具有转移学习的所提出的算法可以实现更好的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号