首页> 美国卫生研究院文献>PLoS Clinical Trials >Research on OpenCL optimization for FPGA deep learning application
【2h】

Research on OpenCL optimization for FPGA deep learning application

机译:FPGA深度学习应用的OpenCL优化研究

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

In recent years, with the development of computer science, deep learning is held as competent enough to solve the problem of inference and learning in high dimensional space. Therefore, it has received unprecedented attention from both the academia and the business community. Compared with CPU/GPU, FPGA has attracted much attention for its high-energy efficiency, short development cycle and reconfigurability in the aspect of deep learning algorithm. However, because of the limited research on OpenCL optimization on FPGA of deep learning algorithms, OpenCL tools and models applied to CPU/GPU cannot be directly used on FPGA. This makes it difficult for software programmers to use FPGA when implementing deep learning algorithms for a rewarding performance. To solve this problem, this paper proposed an OpenCL computational model based on FPGA template architecture to optimize the time-consuming convolution layer in deep learning. The comparison between the program applying the computational model and the corresponding optimization program provided by Xilinx indicates that the former is 8-40 times higher than the latter in terms of performance.
机译:近年来,随着计算机科学的发展,深度学习被认为足以解决高维空间中的推理和学习问题。因此,它受到了学术界和商业界的空前关注。与CPU / GPU相比,FPGA在深度学习算法方面具有高能效,较短的开发周期和可重构性,因此备受关注。但是,由于深度学习算法在FPGA上对OpenCL优化的研究有限,因此,应用于CPU / GPU的OpenCL工具和模型无法直接在FPGA上使用。这使得软件程序员难以在实现深度学习算法以提高性能时使用FPGA。为了解决这个问题,本文提出了一种基于FPGA模板架构的OpenCL计算模型,以优化深度学习中耗时的卷积层。应用计算模型的程序与Xilinx提供的相应优化程序之间的比较表明,在性能方面,前者比后者高8-40倍。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号