...
首页> 外文期刊>ACM transactions on reconfigurable technology and systems >FINN-R: An End-to-End Deep-Learning Framework for Fast Exploration of Quantized Neural Networks
【24h】

FINN-R: An End-to-End Deep-Learning Framework for Fast Exploration of Quantized Neural Networks

机译:FINN-R:快速探索量化神经网络的端到端深度学习框架

获取原文
获取原文并翻译 | 示例
           

摘要

Convolutional Neural Networks have rapidly become the most successful machine-learning algorithm, enabling ubiquitous machine vision and intelligent decisions on even embedded computing systems. While the underlying arithmetic is structurally simple, compute and memory requirements are challenging. One of the promising opportunities is leveraging reduced-precision representations for inputs, activations, and model parameters. The resulting scalability in performance, power efficiency, and storage footprint provides interesting design compromises in exchange for a small reduction in accuracy. FPGAs are ideal for exploiting low-precision inference engines leveraging custom precisions to achieve the required numerical accuracy for a given application. In this article, we describe the second generation of the FINN framework, an end-to-end tool that enables design-space exploration and automates the creation of fully customized inference engines on FPGAs. Given a neural network description, the tool optimizes for given platforms, design targets, and a specific precision. We introduce formalizations of resource cost functions and performance predictions and elaborate on the optimization algorithms. Finally, we evaluate a selection of reduced precision neural networks ranging from CIFAR-10 classifiers to YOLO-based object detection on a range of platforms including PYNQ and AWS F1, demonstrating new unprecedented measured throughput at 50 TOp/s on AWS F1 and 5 TOp/s on embedded devices.
机译:卷积神经网络已迅速成为最成功的机器学习算法,甚至在嵌入式计算系统上也可以实现无处不在的机器视觉和智能决策。尽管底层算法在结构上很简单,但计算和内存需求却充满挑战。一种有前途的机会是利用输入,激活和模型参数的简化表示形式。由此产生的性能,电源效率和存储空间的可扩展性提供了有趣的设计折衷,以换取精度的少量降低。 FPGA是利用定制精度来开发给定应用所需的数值精度的低精度推理引擎的理想选择。在本文中,我们描述了第二代FINN框架,这是一种端到端工具,可用于设计空间探索并自动在FPGA上创建完全定制的推理引擎。给定一个神经网络描述,该工具将针对给定的平台,设计目标和特定的精度进行优化。我们介绍了资源成本函数和性能预测的形式化,并详细说明了优化算法。最后,我们在包括PYNQ和AWS F1在内的一系列平台上评估了从CIFAR-10分类器到基于YOLO的对象检测等一系列降低精度的神经网络,展示了AWS F1和5 TOp上以50 TOp / s的新空前测量吞吐量/ s在嵌入式设备上。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号