首页> 外文会议>IEEE International Symposium on High Performance Computer Architecture >Communication Lower Bound in Convolution Accelerators
【24h】

Communication Lower Bound in Convolution Accelerators

机译:卷积加速器中的通信下限

获取原文

摘要

In current convolutional neural network (CNN) accelerators, communication (i.e., memory access) dominates the energy consumption. This work provides comprehensive analysis and methodologies to minimize the communication for CNN accelerators. For the off-chip communication, we derive the theoretical lower bound for any convolutional layer and propose a dataflow to reach the lower bound. This fundamental problem has never been solved by prior studies. The on-chip communication is minimized based on an elaborate workload and storage mapping scheme. We in addition design a communication-optimal CNN accelerator architecture. Evaluations based on the 65nm technology demonstrate that the proposed architecture nearly reaches the theoretical minimum communication in a three-level memory hierarchy and it is computation dominant. The gap between the energy efficiency of our accelerator and the theoretical best value is only 37-87%.
机译:在当前的卷积神经网络(CNN)加速器中,通信(即,内存访问)主导能耗。这项工作提供了全面的分析和方法,以最大限度地减少CNN加速器的沟通。对于片外通信,我们导出任何卷积层的理论下限,并提出数据流来达到下限。此基本问题从未通过之前的研究解决。基于精心的工作负载和存储映射方案,可以最小化芯片通信。我们此外,我们还设计了一种通信 - 最佳CNN加速器架构。基于65nm技术的评估表明,所提出的架构几乎达到三级内存层级中的理论最小通信,并且它是计算主导。我们加速器的能效与理论最佳价值之间的差距仅为37-87%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号