首页> 美国卫生研究院文献>other >RSTensorFlow: GPU Enabled TensorFlow for Deep Learning on Commodity Android Devices
【2h】

RSTensorFlow: GPU Enabled TensorFlow for Deep Learning on Commodity Android Devices

机译:RSTensorFlow:启用GPU的TensorFlow在商品Android设备上进行深度学习

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Mobile devices have become an essential part of our daily lives. By virtue of both their increasing computing power and the recent progress made in AI, mobile devices evolved to act as intelligent assistants in many tasks rather than a mere way of making phone calls. However, popular and commonly used tools and frameworks for machine intelligence are still lacking the ability to make proper use of the available heterogeneous computing resources on mobile devices. In this paper, we study the benefits of utilizing the heterogeneous (CPU and GPU) computing resources available on commodity android devices while running deep learning models. We leveraged the heterogeneous computing framework RenderScript to accelerate the execution of deep learning models on commodity Android devices. Our system is implemented as an extension to the popular open-source framework TensorFlow. By integrating our acceleration framework tightly into TensorFlow, machine learning engineers can now easily make benefit of the heterogeneous computing resources on mobile devices without the need of any extra tools. We evaluate our system on different android phones models to study the trade-offs of running different neural network operations on the GPU. We also compare the performance of running different models architectures such as convolutional and recurrent neural networks on CPU only vs using heterogeneous computing resources. Our result shows that although GPUs on the phones are capable of offering substantial performance gain in matrix multiplication on mobile devices. Therefore, models that involve multiplication of large matrices can run much faster (approx. 3 times faster in our experiments) due to GPU support.
机译:移动设备已成为我们日常生活的重要组成部分。凭借其不断增强的计算能力和AI方面的最新进展,移动设备已发展为在许多任务中充当智能助手,而不仅仅是打电话的一种方式。但是,流行的和常用的机器智能工具和框架仍然缺乏正确使用移动设备上可用的异构计算资源的能力。在本文中,我们研究了运行深度学习模型时利用商用android设备上可用的异构(CPU和GPU)计算资源的好处。我们利用异构计算框架 RenderScript 加快了在商用Android设备上深度学习模型的执行速度。我们的系统是对流行的开源框架 TensorFlow 的扩展。通过将我们的加速框架紧密集成到 TensorFlow 中,机器学习工程师现在可以轻松地利用移动设备上的异构计算资源,而无需任何其他工具。我们在不同的Android手机模型上评估我们的系统,以研究在GPU上运行不同的神经网络操作的权衡。我们还比较了仅在CPU上与使用异构计算资源时运行不同模型体系结构(例如卷积和循环神经网络)的性能。我们的结果表明,尽管手机上的GPU能够在移动设备上的矩阵乘法中提供可观的性能提升。因此,由于GPU的支持,涉及大型矩阵乘法的模型可以运行得更快(在我们的实验中,快约3倍)。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号