首页> 外文会议>Asia and South Pacific Design Automation Conference >Dynamic Programming Assisted Quantization Approaches for Compressing Normal and Robust DNN Models
【24h】

Dynamic Programming Assisted Quantization Approaches for Compressing Normal and Robust DNN Models

机译:动态编程辅助量化方法,用于压缩正常和强大的DNN模型

获取原文

摘要

In this work, we present effective quantization approaches for compressing the deep neural networks (DNNs). A key ingredient is a novel dynamic programming (DP) based algorithm to obtain the optimal solution of scalar K-means clustering. Based on the approaches with regularization and quantization function, two weight quantization approaches called DPR and DPQ for compressing normal DNNs are proposed respectively. Experiments show that they produce models with higher inference accuracy than recently proposed counterparts while achieving same or larger compression. They are also extended for compressing robust DNNs, and the relevant experiments show 16X compression of the robust ResNet-18 model with less than 3% accuracy drop on both natural and adversarial examples.
机译:在这项工作中,我们提出了用于压缩深度神经网络(DNN)的有效量化方法。关键成分是一种基于新的动态编程(DP)算法,以获得标量k-means聚类的最佳解决方案。基于具有正则化和量化功能的方法,提出了两个称为DPR和DPQ用于压缩正常DNN的重量量化方法。实验表明,它们生产比最近提出的对应物更高推理的型号,同时实现相同或更大的压缩。它们还延长了压缩稳健的DNN,相关实验表明了鲁棒RESET-18型号的16倍压缩,在天然和对抗的例子上具有小于3%的精度下降。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号