【24h】

ParLearning 2018 Invited Talk 2

机译:Parlearning 2018邀请谈判2

获取原文

摘要

The past few years have witnessed a growth in size and computational requirements for training and inference with neural networks. Currently, a common approach to address these requirements is to use a heterogeneous distributed environment with a mixture of hardware devices such as CPUs and GPUs. Importantly, the decision of placing parts of the neural models on devices is often made by human experts based on simple heuristics and intuitions. In this talk, I will present some of our recent efforts on learning to optimize model parallelism for TensorFlow computational graphs. Key to our method is the use of deep reinforcement learning to predict which subsets of operations in a TensorFlow graph should run on which of the available devices. The execution time of the predicted placements is then used as the reward signal to optimize the parameters of the deep model. Our main result is that on important computer vision, language modeling and neural machine translation tasks, our model finds non-trivial ways to parallelise the model that outperform hand-crafted heuristics and traditional algorithmic methods.
机译:过去几年目睹了对神经网络的培训和推论的规模和计算要求的增长。目前,解决这些要求的常见方法是使用异构分布式环境,其中硬件设备如CPU和GPU的混合。重要的是,在设备上放置部分神经模型的决定通常由人类专家基于简单的启发式和直觉。在这次谈话中,我将展示我们最近的一些努力,以优化Tensorflow计算图的模型并行性。我们的方法的关键是使用深度增强学习,以预测Tensorflow图中的哪个操作子集应该在哪个可用设备上运行。然后将预测的放置的执行时间用作奖励信号以优化深模型的参数。我们的主要结果是,在重要的计算机视觉,语言建模和神经机翻译任务上,我们的模型发现非琐碎的方法是并行的,可以平行于表达手工制作的启发式和传统算法方法的模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号