首页> 外文会议>Conference on Ph.D. Research in Microelectronics and Electronics >Improving Deep Learning with a customizable GPU-like FPGA-based accelerator
【24h】

Improving Deep Learning with a customizable GPU-like FPGA-based accelerator

机译:使用可定制的类似GPU的基于FPGA的加速器改善深度学习

获取原文

摘要

An ever increasing number of challenging applications are being approached using Deep Learning, obtaining impressive results in a variety of different domains. However, state-of-the-art accuracy requires deep neural networks with a larger number of layers and a huge number of different filters with millions of weights. GPUand FPGA-based architectures have been proposed as a possible solution for facing this enormous demand of computing resources. In this paper, we investigate the adoption of different architectural features, i.e. SIMD paradigm, multithreading, and non-coherent on-chip memory for Deep Learning oriented FPGA-based accelerator designs. Experimental results on a Xilinx Virtex-7 FPGA show that the SIMD paradigm and multithreading can lead to an improvement in the execution time up to 5× and 3.5×, respectively. A further enhancement up to 1.75× can be obtained using a non-coherent on-chip memory.
机译:深度学习正在处理越来越多具有挑战性的应用程序,从而在各种不同领域中均获得了令人印象深刻的结果。但是,最新的准确性要求深度神经网络具有更多的层数以及大量具有数百万个权重的不同过滤器。已经提出了基于GPU和FPGA的体系结构作为应对这种巨大的计算资源需求的可能解决方案。在本文中,我们研究了针对面向深度学习的基于FPGA的加速器设计采用不同的架构功能(即SIMD范例,多线程和非一致性片上存储器)的情况。在Xilinx Virtex-7 FPGA上的实验结果表明,SIMD范例和多线程可以分别将执行时间缩短多达5倍和3.5倍。使用非相干的片上存储器可以获得高达1.75x的进一步增强。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号