首页> 外文会议>IEEE Conference on Computer Vision and Pattern Recognition Workshops >Binarized Convolutional Neural Networks with Separable Filters for Efficient Hardware Acceleration
【24h】

Binarized Convolutional Neural Networks with Separable Filters for Efficient Hardware Acceleration

机译:具有可分离滤波器的二值化卷积神经网络,用于有效的硬件加速度

获取原文

摘要

State-of-the-art convolutional neural networks are enormously costly in both compute and memory, demanding massively parallel GPUs for execution. Such networks strain the computational capabilities and energy available to embedded and mobile processing platforms, restricting their use in many important applications. In this paper, we propose BCNN with Separable Filters (BCNNw/SF), which applies Singular Value Decomposition (SVD) on BCNN kernels to further reduce computational and storage complexity. We provide a closed form of the gradient over SVD to calculate the exact gradient with respect to every binarized weight in backward propagation. We verify BCNNw/SF on the MNIST, CIFAR-10, and SVHN datasets, and implement an accelerator for CIFAR10 on FPGA hardware. Our BCNNw/SF accelerator realizes memory savings of 17% and execution time reduction of 31.3% compared to BCNN with only minor accuracy sacrifices.
机译:在计算和记忆中,最先进的卷积神经网络在两个计算和内存中都是昂贵的,要求大量平行的GPU进行执行。这种网络应变为嵌入式和移动处理平台提供的计算能力和能量,限制它们在许多重要应用中的使用。在本文中,我们提出了具有可分离滤波器(BCNNW / SF)的BCNN,其在​​BCNN内核上应用奇异值分解(SVD),以进一步降低计算和存储复杂性。我们在SVD提供梯度的闭合形式,以在向后传播中计算相对于每个二值化重量的精确梯度。我们在Mnist,CiFar-10和SVHN数据集上验证BCNNW / SF,并在FPGA硬件上实施CIFAR10的加速器。我们的BCNNW / SF Accelerator与BCNN相比,4 %的内存节省为17 %并执行时间减少31.3 %,只有轻微的精度牺牲。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号