首页> 外文会议>Asia and South Pacific Design Automation Conference >BHNN: A memory-efficient accelerator for compressing deep neural networks with blocked hashing techniques
【24h】

BHNN: A memory-efficient accelerator for compressing deep neural networks with blocked hashing techniques

机译:BHNN:一种内存有效的加速器,用于使用分块哈希技术压缩深度神经网络

获取原文

摘要

In this paper, we propose a novel algorithm for compressing neural networks to reduce the memory requirements by using blocked hashing techniques. By adding blocked constraints on top of the conventional hashing technique, the test error rate is maintained while the spatial locality for the computations is preserved. Using this scheme, the synaptic connections are compressed by at least an order (10×) compared with the plain neural network with virtually no prediction accuracy loss. Compared with other compression techniques, the proposed algorithm achieves the best performance in the heavy compression regions. The blocked hashing techniques are also hardware friendly, of which the memory hierarchy of the hardware architecture can be efficiently implemented. To demonstrate the hardware efficiency, we implement the hardware architecture of the deep neural networks using the proposed blocked hashing techniques on a Xilinx Virtex-7 FPGA board. With a hardware parallelism of 32, the accelerator achieves a speed-up of 22× over the CPU, and 3~5× over the GPU in the inference phase.
机译:在本文中,我们提出了一种新的算法,用于压缩神经网络,以通过使用分块哈希技术来减少内存需求。通过在常规哈希技术的基础上添加阻塞约束,可以保持测试错误率,同时保留计算的空间局部性。使用该方案,与普通神经网络相比,突触连接至少压缩了一个顺序(10倍),几乎没有预测准确性的损失。与其他压缩技术相比,该算法在重压缩区域达到了最佳性能。分块哈希技术也是硬件友好的,可以有效地实现硬件体系结构的内存层次结构。为了演示硬件效率,我们在Xilinx Virtex-7 FPGA板上使用提出的分块哈希技术实现了深度神经网络的硬件架构。硬件并行度为32,在推理阶段,加速器在CPU上的速度提高了22倍,在GPU上的速度提高了3〜5倍。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号