首页> 外文会议>International conference on neural information processing >Compressing Low Precision Deep Neural Networks Using Sparsity-Induced Regularizat ion in Ternary Networks
【24h】

Compressing Low Precision Deep Neural Networks Using Sparsity-Induced Regularizat ion in Ternary Networks

机译:三元网络中稀疏性引起的正则化压缩低精度深度神经网络

获取原文

摘要

A low precision deep neural network training technique for producing sparse, ternary neural networks is presented. The technique incorporates hardware implementation costs during training to achieve significant model compression for inference. Training involves three stages: network training using L2 regularization and a quantization threshold regularizer, quantization pruning, and finally retraining. Resulting networks achieve improved accuracy, reduced memory footprint and reduced computational complexity compared with conventional methods, on MNIST and CIFAR10 datasets. Our networks are up to 98% sparse and 5 & 11 times smaller than equivalent binary and ternary models, translating to significant resource and speed benefits for hardware implementations.
机译:提出了一种用于生成稀疏三元神经网络的低精度深度神经网络训练技术。该技术在训练过程中并入了硬件实施成本,以实现显着的模型压缩以进行推理。训练包括三个阶段:使用L2正则化和量化阈值正则化器进行网络训练,量化修剪以及最后重新训练。与传统方法相比,结果网络在MNIST和CIFAR10数据集上实现了更高的准确性,更少的内存占用以及更低的计算复杂性。我们的网络稀疏度高达98%,比等效的二进制和三进制模型小5到11倍,这为硬件实现带来了显着的资源和速度优势。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号