首页> 外文会议>ISC High Performance Conference >Distributed Deep Reinforcement Learning: Learn How to Play Atari Games in 21 minutes
【24h】

Distributed Deep Reinforcement Learning: Learn How to Play Atari Games in 21 minutes

机译:分布式深度加强学习:学习如何在21分钟内玩Atari游戏

获取原文

摘要

We present a study in Distributed Deep Reinforcement Learning (DDRL) focused on scalability of a state-of-the-art Deep Reinforcement Learning algorithm known as Batch Asynchronous Advantage Actor-Critic, (BA3C). We show that using the Adam optimization algorithm with a batch size of up to 2048 is a viable choice for carrying out large scale machine learning computations. This, combined with careful reexamination of the optimizer's hyperparameters, using synchronous training on the node level (while keeping the local, single node part of the algorithm asynchronous) and minimizing the model's memory footprint, allowed us to achieve linear scaling for up to 64 CPU nodes. This corresponds to a training time of 21 min on 768 CPU cores, as opposed to the 10 h required when using a single node with 24 cores achieved by a baseline single-node implementation.
机译:我们在分布式深度加强学习(DDRL)中展示了专注于称为批量异步优势演员 - 评论家(BA3C)的最先进的深增强学习算法的可扩展性。我们表明,使用批量大小最多可达2048次的ADAM优化算法是进行大规模机器学习计算的可行选择。这将仔细复制优化器的超参数,使用节点级别的同步训练(在保持算法的本地,单节点部分异步)并最大限度地减少模型的内存占用空间,使我们能够实现最多64个CPU的线性缩放节点。这对应于768 CPU核心21分钟的训练时间,而不是使用由基线单节点实现实现的24个核心所需的10小时。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号