首页> 外文会议>National Conference with International Participation >Using of Bfloat16 Format in Deep Learning Embedded Accelerators based on FPGA with Limited Quantity of Dedicated Multipliers
【24h】

Using of Bfloat16 Format in Deep Learning Embedded Accelerators based on FPGA with Limited Quantity of Dedicated Multipliers

机译:基于FPGA的深度学习嵌入式加速器中的Bfloat16格式,具有限量专用乘法器

获取原文

摘要

The hardware base of Deep Learning Neural Network (DLNN) realization methods are remote cloud services, Graphical Processing Units (GPU) and Field Programmable Gate Arrays (FPGA). The one of the main differences between FPGA devices is important for DLNN realization is quantity of dedicated multipliers in DSP blocks. In this article a method for optimization based on bfloat16 data format useful for FPGA devices with small quantities of DSP blocks is described.
机译:深度学习神经网络(DLNN)实现方法的硬件基础是远程云服务,图形处理单元(GPU)和现场可编程门阵列(FPGA)。 FPGA设备之间的主要差异对于DLNN实现是重要的,这是DSP块中专用乘法器的数量。在本文中,描述了一种基于对具有少量DSP块的FPGA设备的BFLOAT16数据格式进行优化方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号