...
首页> 外文期刊>IEEE computational intelligence magazine >Improving RTS Game AI by Supervised Policy Learning, Tactical Search, and Deep Reinforcement Learning
【24h】

Improving RTS Game AI by Supervised Policy Learning, Tactical Search, and Deep Reinforcement Learning

机译:通过监督策略学习,战术搜索和深度强化学习来改善RTS Game AI

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Constructing strong AI systems for video games is difficult due to enormous state and action spaces and the lack of good state evaluation functions and high-level action abstractions. In spite of recent research progress in popular video game genres such as Atari 2600 console games and multiplayer online battle arena (MOBA) games, to this day strong human players can still defeat the best AI systems in adversarial video games. In this paper, we propose to use a deep Convolutional Neural Network (CNN) to select among a limited set of abstract action choices in Real-Time Strategy (RTS) games, and to utilize the remaining computation time for game tree search to improve low-level tactics. The CNN is trained by supervised learning on game states labeled by Puppet Search, a strategic search algorithm that uses action abstractions. Replacing Puppet Search by a CNN frees up time that can be used for improving units' tactical behavior while executing the strategic plan. Experiments in the mu RTS game show that the combined algorithm results in higher win-rates than either of its two independent components and other state-of-the-art mu RTS agents. We then present a case-study that investigates how deep Reinforcement Learning (RL) can be used in modern video games, such as Total War: Warhammer, to improve tactical multi-agent AI modules. We use popular RL algorithms such as Deep-Q Networks (DQN) and Asynchronous AdvantageActor Critic (A3C), basic network architectures and minimal hyper-parameter tuning to learn complex cooperative behaviors that defeat the highest difficulty built-in AI in mediumscale scenarios.
机译:由于巨大的状态和动作空间以及缺乏良好的状态评估功能和高级动作抽象,为视频游戏构建强大的AI系统非常困难。尽管最近流行的视频游戏类型(如Atari 2600主机游戏和多人在线战斗竞技场(MOBA)游戏)取得了研究进展,但迄今为止,强大的人类玩家仍然可以击败对抗性视频游戏中最好的AI系统。在本文中,我们建议使用深度卷积神经网络(CNN)在实时策略(RTS)游戏中的一组有限的抽象动作选择中进行选择,并利用剩余的计算时间进行游戏树搜索以降低级战术。 CNN通过监督学习来训练以Puppet Search标记的游戏状态,Puppet Search是一种使用动作抽象的战略搜索算法。用CNN取代人偶搜索可以节省时间,这些时间可用于在执行战略计划时改善部队的战术行为。在mu RTS游戏中进行的实验表明,结合后的算法比其两个独立组件和其他最新的mu RTS代理中的任何一个都具有更高的获胜率。然后,我们提供一个案例研究,研究如何在诸如Total War:Warhammer之类的现代视频游戏中使用深度强化学习(RL)来改进战术多主体AI模块。我们使用诸如Deep-Q Networks(DQN)和Asynchronous AdvantageActor Critic(A3C)等流行的RL算法,基本的网络体系结构和最小化的超参数调整来学习复杂的协作行为,以克服中等规模场景中内置AI的最高难度。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号