首页> 外文会议>IEEE International Symposium on Circuits and Systems >High Utilization Energy-Aware Real-Time Inference Deep Convolutional Neural Network Accelerator
【24h】

High Utilization Energy-Aware Real-Time Inference Deep Convolutional Neural Network Accelerator

机译:高利用能量感知实时推断深卷积神经网络加速器

获取原文
获取外文期刊封面目录资料

摘要

Deep convolution Neural Network (DCNN) has been widely used in computer vision tasks. However, for edge device, even then inference has too large computational complexity and data access amount. Due to the mentioned shortcomings, the inference latency of state-of-the-art models are still impractical for real-world applications. In this paper, we proposed a high utilization energy-aware real-time inference deep convolutional neural network accelerator, which outperforms the current accelerators. First, we use 1×1 size convolution kernels as the smallest unit of the computing unit. And we design suitable computing unit for different models based on the requirement of each model. Second, we use Reuse Feature SRAM to store the output of current layer in the chip and use as the input of the next layer. Moreover, we import Output Reuse Strategy and Ring Stream Data flow not only to expand the reuse rate of data in the chip but to reduce the amount of data exchange between chips and DRAM. Finally, we present On-fly Pooling Module to let the calculation of the Pooling layer to be completed directly in the chip. With the aid of the proposed method in this paper, the implemented CNN acceleration chip has extreme high hardware utilization rate. We reduce a generous amount of data transfer on the specific module, ECNN [1]. Compared to the methods without reuse strategy, we can reduce 533 times of data access amount. At the same time, we have enough computing power to perform real-time execution of the existing image classification model, VGG16 [2] and MobileNet [3]. Compared with the design in [4], we can speed up 7.52 times and have 1.92x energy efficiency.
机译:深度卷积神经网络(DCNN)已广泛应用于计算机视觉任务。但是,对于边缘设备,即使是推理也具有过大的计算复杂性和数据访问量。由于提到的缺点,最先进的模型的推理延迟对于真实的应用仍然是不切实际的。在本文中,我们提出了高利用能量感知的实时推理深卷积神经网络加速器,其优于当前的加速器。首先,我们使用1×1尺寸卷积内核作为计算单元的最小单位。我们根据每个模型的要求,为不同模型设计合适的计算单元。其次,我们使用重用功能SRAM将电流层的输出存储在芯片中并用作下一层的输入。此外,我们导入输出重用策略和环流数据流不仅要扩展芯片中的数据的重用率,而且还可以减少芯片和DRAM之间的数据交换量。最后,我们展示了可在飞行池模块,让池池层的计算直接在芯片中完成。借助于本文所提出的方法,实施的CNN加速芯片具有极高的硬件利用率。我们减少了特定模块上的大量数据传输,ECNN [1]。与没有重用策略的方法相比,我们可以减少533次数据访问量。同时,我们有足够的计算能力来执行现有图像分类模型的实时执行,VGG16 [2]和MobileNet [3]。与[4]中的设计相比,我们可以加快7.52次,并具有1.92倍的能量效率。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号