首页> 外文期刊>Journal of circuits, systems and computers >Optimizing Convolutional Neural Network Accelerator on Low-Cost FPGA
【24h】

Optimizing Convolutional Neural Network Accelerator on Low-Cost FPGA

机译:优化卷积神经网络加速器低成本FPGA

获取原文
获取原文并翻译 | 示例
           

摘要

Convolutional neural network (CNN) is one of the most promising algorithms that outweighs other traditional methods in terms of accuracy in classification tasks. However, several CNNs, such as VGG, demand a huge computation in convolutional layers. Many accelerators implemented on powerful FPGAs have been introduced to address the problems. In this paper, we present a VGG-based accelerator which is optimized for a low-cost FPGA. In order to optimize the FPGA resource of logic element and memory, we propose a dedicated input buffer that maximizes the data reuse. In addition, we design a low resource processing engine with the optimal number of Multiply Accumulate (MAC) units. In the experiments, we use VGG16 model for inference to evaluate the performance of our accelerator and achieve a throughput of 38.8GOPS at a clock speed of 150MHz on Intel Cyclone V SX SoC. The experimental results show that our design is better than previous works in terms of resource efficiency.
机译:卷积神经网络(CNN)是最有前途的算法之一,在分类任务中的准确性方面超过了其他传统方法。 然而,几个CNN,例如VGG,需要在卷积层中计算巨大的计算。 已经引入了许多在强大的FPGA上实施的加速器来解决问题。 在本文中,我们介绍了一种基于VGG的加速器,其针对低成本FPGA进行了优化。 为了优化逻辑元素和内存的FPGA资源,我们提出了一个专用输入缓冲区,可最大化数据重用。 此外,我们设计了具有最佳累积(MAC)单元的最佳数量的低资源处理引擎。 在实验中,我们使用VGG16模型来推理,以评估我们的加速器的性能,并在英特尔Cyclone V SX SoC上以150MHz的时钟速度实现38.8GGOP的吞吐量。 实验结果表明,在资源效率方面,我们的设计优于以前的工作。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号