首页> 外文会议>Design, Automation and Test in Europe Conference and Exhibition >Deeper Weight Pruning without Accuracy Loss in Deep Neural Networks
【24h】

Deeper Weight Pruning without Accuracy Loss in Deep Neural Networks

机译:在深度神经网络中进行更深的权重修剪而没有精度损失

获取原文

摘要

This work overcomes the inherent limitation of the bit-level weight pruning, that is, the maximal computation speedup is bounded by the total number of non-zero bits of the weights and the bound is invariably considered "uncontrollable" (i.e., constant) for the neural network to be pruned. Precisely, this work, based on the canonical signed digit (CSD) encoding, (1) proposes a transformation technique which converts the two’s complement representation of every weight into a set of CSD representations of the minimal or near-minimal number of essential (i.e., non-zero) bits, (2) formulates the problem of selecting CSD representations of weights that maximize the parallelism of bit-level multiplication on the weights into a multi-objective shortest path problem and solves it efficiently using an approximation algorithm, and (3) proposes a supporting novel acceleration architecture with no additional inclusion of non-trivial hardware. Through experiments, it is shown that our proposed approach reduces the number of essential bits by 69% on AlexNet and 74% on VGG-16, by which our accelerator reduces the inference computation time by 47% on AlexNet and 50% on VGG-16 over the conventional bit-level weight pruning.
机译:这项工作克服了位级权重修剪的固有局限性,即最大的计算速度受到权重的非零位总数的限制,并且对于被修剪的神经网络。准确地说,这项工作基于规范符号数字(CSD)编码,(1)提出了一种转换技术,可以将每个权重的二进制补码表示形式转换为一组最小或接近最小基本数字的CSD表示形式(即(非零)位,(2)提出了选择权重的CSD表示法的问题,这些表示法将权重上的位级乘法的并行性最大化为多目标最短路径问题,并使用近似算法有效地解决了该问题,并且( 3)提出了一种支持性的新颖加速架构,其中没有额外包含非平凡的硬件。通过实验表明,我们提出的方法将AlexNet上的基本位数减少了69%,将VGG-16上的基本位数减少了74%,这样,我们的加速器将AlexNet上的推理计算时间减少了47%,将VGG-16上的推理计算时间减少了50%。超过了传统的位级权重修剪。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号