首页> 外文会议>IEEE International Conference on Autonomous Robot Systems and Competitions >Competitive Deep Reinforcement Learning over a Pokémon Battling Simulator
【24h】

Competitive Deep Reinforcement Learning over a Pokémon Battling Simulator

机译:通过神奇宝贝战斗模拟器进行的竞争性深度强化学习

获取原文

摘要

Pokémon is one of the most popular video games in the world, and recent interest has appeared in Pokémon battling as a testbed for AI challenges. This is due to Pokémon battling showing interesting properties which contrast with current AI challenges over other video games. To this end, we implement a Pokémon Battle Environment, which preserves many of the core elements of Pokémon battling, and allows researchers to test isolated learning objectives. Our approach focuses on type advantage in Pokémon battles and on the advantages of delayed rewards through switching, which is considered core strategies for any Pokémon battle. As a competitive multi-agent environment, it has a partially-observable, high-dimensional, and continuous state-space, adheres to the Gym de facto standard reinforcement learning interface, and is performance-oriented, achieving thousands of interactions per second in commodity hardware. We determine whether deep competitive reinforcement learning algorithms, WPLθ and GIGAθ, can learn successful policies in this environment. Both converge to rational and effective strategies, and GIGAθ shows faster convergence, obtaining a 100% win-rate in a disadvantageous test scenario.
机译:神奇宝贝是世界上最受欢迎的视频游戏之一,最近对于神奇宝贝战斗的兴趣也逐渐显现出来,以此作为应对AI挑战的试验台。这是由于《神奇宝贝》的战斗表现出了有趣的特性,这与当前AI挑战其他视频游戏形成了鲜明的对比。为此,我们实施了一个神奇宝贝战斗环境,该环境保留了神奇宝贝战斗的许多核心要素,并允许研究人员测试孤立的学习目标。我们的方法侧重于神奇宝贝战斗中的类型优势,以及通过切换获得延迟奖励的优势,这被认为是任何神奇宝贝战斗的核心策略。作为竞争性的多主体环境,它具有部分可观察的,高维且连续的状态空间,并遵循Gym de facto标准的强化学习界面,并且以性能为导向,每秒可实现数千次商品交互硬件。我们确定深度竞争性强化学习算法WPLθ和GIGAθ是否可以在这种环境下学习成功的策略。两者都收敛于合理有效的策略,并且GIGAθ显示出更快的收敛速度,在不利的测试场景中获得了100%的获胜率。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号