首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Device Placement Optimization with Reinforcement Learning
【24h】

Device Placement Optimization with Reinforcement Learning

机译:通过强化学习优化设备放置

获取原文
           

摘要

The past few years have witnessed a growth in size and computational requirements for training and inference with neural networks. Currently, a common approach to address these requirements is to use a heterogeneous distributed environment with a mixture of hardware devices such as CPUs and GPUs. Importantly, the decision of placing parts of the neural models on devices is often made by human experts based on simple heuristics and intuitions. In this paper, we propose a method which learns to optimize device placement for TensorFlow computational graphs. Key to our method is the use of a sequence-to-sequence model to predict which subsets of operations in a TensorFlow graph should run on which of the available devices. The execution time of the predicted placements is then used as the reward signal to optimize the parameters of the sequence-to-sequence model. Our main result is that on Inception-V3 for ImageNet classification, and on RNN LSTM, for language modeling and neural machine translation, our model finds non-trivial device placements that outperform hand-crafted heuristics and traditional algo-rithmic methods.
机译:过去几年见证了神经网络训练和推理的规模和计算需求的增长。当前,解决这些要求的常用方法是使用异构分布式环境,并混合使用诸如CPU和GPU之类的硬件设备。重要的是,通常由人类专家根据简单的试探法和直觉来做出将部分神经模型放置在设备上的决定。在本文中,我们提出了一种方法来学习优化TensorFlow计算图的设备位置。我们方法的关键是使用序列到序列模型来预测TensorFlow图中的哪些操作子集应在哪些可用设备上运行。然后,将预测放置的执行时间用作奖励信号,以优化序列到序列模型的参数。我们的主要结果是,在用于ImageNet分类的Inception-V3以及用于语言建模和神经机器翻译的RNN LSTM上,我们的模型发现了非平凡的设备布局,其性能优于手工启发式算法和传统的算法算法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号