首页> 外文会议>International Conference on Neural Information Processing >Compressing Low Precision Deep Neural Networks Using Sparsity-Induced Regularization in Ternary Networks
【24h】

Compressing Low Precision Deep Neural Networks Using Sparsity-Induced Regularization in Ternary Networks

机译:在三元网络中使用稀疏性致正则化压缩低精度深神经网络

获取原文

摘要

A low precision deep neural network training technique for producing sparse, ternary neural networks is presented. The technique incorporates hardware implementation costs during training to achieve significant model compression for inference. Training involves three stages: network training using L2 regularization and a quantization threshold regularizer, quantization pruning, and finally retraining. Resulting networks achieve improved accuracy, reduced memory foot-print and reduced computational complexity compared with conventional methods, on MNIST and CIFAR10 datasets. Our networks are up to 98% sparse and 5 & 11 times smaller than equivalent binary and ternary models, translating to significant resource and speed benefits for hardware implementations.
机译:提出了一种用于生产稀疏,三元神经网络的低精度深度神经网络培训技术。该技术在训练期间包含硬件实施成本,以实现推理的显着模型压缩。培训涉及三个阶段:使用L2正则化的网络训练和量化阈值规范器,量化修剪,最后再培训。与传统方法相比,由MniSt和CiFar10数据集相比,所产生的网络实现了更高的准确性,降低了内存脚印和降低的计算复杂性。我们的网络稀疏高达98%,比同等二进制和三元模型小5倍和11倍,转化为硬件实现的大量资源和速度优势。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号