首页> 外文会议>Conference on Neural Information Processing Systems >AutoPrune: Automatic Network Pruning by Regularizing Auxiliary Parameters
【24h】

AutoPrune: Automatic Network Pruning by Regularizing Auxiliary Parameters

机译:Autoprune:通过正规化辅助参数自动网络修剪

获取原文
获取外文期刊封面目录资料

摘要

Reducing the model redundancy is an important task to deploy complex deep learning models to resource-limited or time-sensitive devices. Directly regularizing or modifying weight values makes pruning procedure less robust and sensitive to the choice of hyperparameters, and it also requires prior knowledge to tune different hyperparameters for different models. To build a better generalized and easy-to-use pruning method, we propose AutoPrune, which prunes the network through optimizing a set of trainable auxiliary parameters instead of original weights. The instability and noise during training on auxiliary parameters will not directly affect weight values, which makes pruning process more robust to noise and less sensitive to hyperparameters. Moreover, we design gradient update rules for auxiliary parameters to keep them consistent with pruning tasks. Our method can automatically eliminate network redundancy with recoverability, relieving the complicated prior knowledge required to design thresholding functions, and reducing the time for trial and error. We evaluate our method with LeNet and VGG-like on MNIST and CIFAR-10 datasets, and with AlexNet, ResNet and MobileNet on ImageNet to establish the scalability of our work. Results show that our model achieves state-of-the-art sparsity, e.g. 7%, 23% FLOPs and 310x, 75x compression ratio for LeNet5 and VGG-like structure without accuracy drop, and 200M and 100M FLOPs for MobileNet V2 with accuracy 73.32% and 66.83% respectively.
机译:降低模型冗余是将复杂的深度学习模型部署到资源限制或时间敏感设备的重要任务。直接正常化或修改权重值使修剪过程更强大,对超级参数的选择较不稳定,并且还需要先前的知识来调整不同模型的不同超参数。为了构建更好的广义和易于使用的修剪方法,我们提出了自动缓存,通过优化一系列培训辅助参数而不是原始重量来修剪网络。辅助参数训练期间的不稳定性和噪声不会直接影响重量值,这使修剪过程更加强大地对噪声和对超公数敏感。此外,我们设计辅助参数的渐变更新规则,以保持与修剪任务一致。我们的方法可以通过可恢复性自动消除网络冗余,从而减轻设计阈值函数所需的复杂现有知识,并减少试验和错误的时间。我们在MNIST和CIFAR-10数据集上使用Lenet和VGG-Like评估我们的方法,以及AlexNet,Reset和MobileNet上的Imagenet,以确定我们工作的可扩展性。结果表明,我们的型号实现了最先进的稀疏性,例如, LENET5和VGG样结构的7%,23%拖鞋和310倍,75倍压缩比,无需精确下降,为MOBILENET v2的200m和100m拖鞋,精度分别为73.32%和66.83%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号