首页> 外文会议>2019 56th ACM/IEEE Design Automation Conference >FLightNNs: Lightweight Quantized Deep Neural Networks for Fast and Accurate Inference
【24h】

FLightNNs: Lightweight Quantized Deep Neural Networks for Fast and Accurate Inference

机译:FLightNNs:轻量化的深度神经网络,可进行快速,准确的推断

获取原文
获取原文并翻译 | 示例

摘要

To improve the throughput and energy efficiency of Deep Neural Networks (DNNs) on customized hardware, lightweight neural networks constrain the weights of DNNs to be a limited combination (denoted as k $in$ {1, 2}) of powers of 2. In such networks, the multiply-accumulate operation can be replaced with a single shift operation, or two shifts and an add operation. To provide even more design flexibility, the k for each convolutional filter can be optimally chosen instead of being fixed for every filter. In this paper, we formulate the selection of k to be differentiable, and describe model training for determining k-based weights on a per-filter basis. Over 46 FPGA-design experiments involvmg eight configurations and four data sets reveal that lightweight neural networks with a flexible k value (dubbed FLightNNs) fully utilize the hardware resources on Field Programmable Gate Arrays (FPGAs), our experimental results show that FLightNNs can achieve $2imes$ speedup when compared to lightweight NNs with k = 2, with only 0.1% accuracy degradation. Compared to a 4-bit fixed-point quantization, FLightNNs achieve higher accuracy and up to $2imes$ inference speedup, due to their lightweight shift operations. In adchtion, our experiments also demonstrate that FLightNNs can achieve higher computational energy efficiency for ASIC implementation.CCS CONCEPTS• Computing methodologies $ightarrow$ Machine learning; • Hardware$ightarrow$Electronic design automation.
机译:为了提高自定义硬件上的深度神经网络(DNN)的吞吐量和能效,轻型神经网络将DNN的权重限制为2的幂的有限组合(表示为k $ \ in $ {1,2})在这样的网络中,乘-累加运算可以用单移位运算或两个移位加法运算来代替。为了提供更大的设计灵活性,每个卷积滤波器的k可以最佳选择,而不是每个滤波器都固定。在本文中,我们将k的选择公式化为可微的,并描述了基于每个过滤器确定k权重的模型训练。超过46个FPGA设计实验涉及8种配置和4个数据集,结果表明,具有灵活k值的轻型神经网络(称为FLightNNs)充分利用了现场可编程门阵列(FPGA)上的硬件资源,我们的实验结果表明FLightNNs可以达到2美元与k = 2的轻量级NN相比,提高了\\ times $的速度,而精度下降仅为0.1%。与4位定点量化相比,由于FLightNN的轻量级移位运算,它们实现了更高的准确性,并且推理速度提高了2倍。此外,我们的实验还证明,FLightNNs可以为ASIC实现实现更高的计算能效。CCS概念•计算方法•硬件\\ rightarrow $电子设计自动化。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号