首页> 外文会议>IEEE International Conference on Acoustics, Speech and Signal Processing >Lance: Efficient Low-Precision Quantized Winograd Convolution for Neural Networks Based on Graphics Processing Units
【24h】

Lance: Efficient Low-Precision Quantized Winograd Convolution for Neural Networks Based on Graphics Processing Units

机译:Lance:基于图形处理单元的神经网络高效低精度量化Winograd卷积

获取原文

摘要

Accelerating deep convolutional neural networks has become an active topic and sparked an interest in academia and industry. In this paper, we propose an efficient low-precision quan-tized Winograd convolution algorithm, called LANCE, which combines the advantages of fast convolution and quantization techniques. By embedding linear quantization operations into the Winograd-domain, the fast convolution can be performed efficiently under low-precision computation on graphics processing units. We test neural network models with LANCE on representative image classification datasets, including SVHN, CIFAR, and ImageNet. The experimental results show that our 8-bit quantized Winograd convolution improves the performance by up to 2.40× over the full-precision convolution with trivial accuracy loss.
机译:加速深度卷积神经网络已成为一个活跃的话题,并引起了学术界和工业界的兴趣。在本文中,我们提出了一种有效的低精度量化Winograd量化卷积算法,称为LANCE,它结合了快速卷积和量化技术的优点。通过将线性量化操作嵌入Winograd域,可以在图形处理单元上以低精度计算有效地执行快速卷积。我们在代表性图像分类数据集(包括SVHN,CIFAR和ImageNet)上使用LANCE测试神经网络模型。实验结果表明,与全精度卷积相比,我们的8位量化Winograd卷积将性能提高了2.40倍,而精度却有所降低。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号