首页> 外文会议>Design, Automation and Test in Europe Conference and Exhibition >Towards Design Space Exploration and Optimization of Fast Algorithms for Convolutional Neural Networks (CNNs) on FPGAs
【24h】

Towards Design Space Exploration and Optimization of Fast Algorithms for Convolutional Neural Networks (CNNs) on FPGAs

机译:FPGA上卷积神经网络(CNN)的设计空间探索和快速算法优化

获取原文

摘要

Convolutional Neural Networks (CNNs) have gained widespread popularity in the field of computer vision and image processing. Due to huge computational requirements of CNNs, dedicated hardware-based implementations are being explored to improve their performance. Hardware platforms such as Field Programmable Gate Arrays (FPGAs) are widely being used to design parallel architectures for this purpose. In this paper, we analyze Winograd minimal filtering or fast convolution algorithms to reduce the arithmetic complexity of convolutional layers of CNNs. We explore a complex design space to find the sets of parameters that result in improved throughput and power-efficiency. We also design a pipelined and parallel Winograd convolution engine that improves the throughput and power-efficiency while reducing the computational complexity of the overall system. Our proposed designs show up to 4.75× and 1.44× improvements in throughput and power-efficiency, respectively, in comparison to the state-of-the-art design while using approximately 2.67× more multipliers. Furthermore, we obtain savings of up to 53.6% in logic resources compared with the state-of-the-art implementation.
机译:卷积神经网络(CNN)在计算机视觉和图像处理领域获得了广泛的普及。由于CNN的巨大计算需求,正在探索基于硬件的专用实现以改善其性能。为此,广泛使用诸如现场可编程门阵列(FPGA)之类的硬件平台来设计并行体系结构。在本文中,我们分析了Winograd最小滤波或快速卷积算法,以减少CNN卷积层的算术复杂度。我们探索了一个复杂的设计空间,以找到可提高吞吐量和功率效率的参数集。我们还设计了流水线和并行的Winograd卷积引擎,该引擎提高了吞吐量和功率效率,同时降低了整个系统的计算复杂性。与最新设计相比,我们建议的设计在吞吐量和功率效率方面分别提高了4.75倍和1.44倍,同时使用了大约2.67倍的乘法器。此外,与最新的实现方式相比,我们节省了多达53.6%的逻辑资源。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号