首页> 外文会议>International Conference on Field-Programmable Technology >INFER: INterFerence-aware Estimation of Runtime for Concurrent CNN Execution on DPUs
【24h】

INFER: INterFerence-aware Estimation of Runtime for Concurrent CNN Execution on DPUs

机译:推断:DPU上并发CNN执行运行时的干扰感知估计

获取原文
获取外文期刊封面目录资料

摘要

Deep Learning Processor Unit (DPU) from XILINX is among the numerous accelerators that have been proposed to speed up the execution of Convolutional Neural Networks (CNNs) on embedded platforms. DPUs are available in different configurable sizes and can execute any given CNN. Neural network researchers are also rapidly bringing out newer CNN algorithms with improved performance (typically higher prediction accuracy) with a trade-off in size or energy consumption for embedded applications. To enable quick evaluation of choices among evolving CNN algorithms and accelerator configurations, we propose INFER (INterFerence-aware Estimation of Runtime). INFER is a framework to estimate the execution time of any CNN on a given size of DPU without actual implementation. Further, current FPGA platforms are capable of implementing multiple DPUs whereas many applications consist of multiple sub-tasks with each requiring separate and/or different CNNs. In such scenarios of concurrent use of multiple DPUs on an FPGA, INFER is also capable of estimating the additional time taken for execution due to the sharing of memory bandwidth. Our evaluation on various mixes of 16 standard CNNs and eight configurations of DPU shows that INFER has an average prediction error of 6.6%, which can be useful for design space exploration as well as scheduling in multi-DPU platforms.
机译:来自Xilinx的深度学习处理器单元(DPU)是众多加速器,已经提出加快嵌入式平台上的卷积神经网络(CNNS)的执行。 DPU以不同的可配置尺寸提供,可以执行任何给定的CNN。神经网络研究人员还迅速推出了具有改进的性能(通常更高的预测精度)的新型CNN算法,其嵌入式应用的尺寸或能耗的折衷。为了在不断变化的CNN算法和加速器配置中可以快速评估选择,我们提出推断(运行时的干扰感知估计)。推断是一个框架,用于在没有实际实现的情况下估计在给定大小的DPU大小的任何CNN的执行时间。此外,当前的FPGA平台能够实现多个DPU,而许多应用程序由多个子任务组成,其中每个需要单独和/或不同的CNN。在FPGA上同时使用多个DPU的这种情况下,推断也能够估计由于内存带宽的共享而执行的额外时间。我们对16个标准CNN的各种混合物的评估和DPU的8个配置显示推断的平均预测误差为6.6%,这对于设计空间探索以及在多DPU平台中的调度非常有用。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号