首页> 外文会议>ACM/IEEE Design Automation Conference >HeadStart: Enforcing Optimal Inceptions in Pruning Deep Neural Networks for Efficient Inference on GPGPUs
【24h】

HeadStart: Enforcing Optimal Inceptions in Pruning Deep Neural Networks for Efficient Inference on GPGPUs

机译:Headstart:强制修剪深度神经网络中的最佳初始化,以获得GPGPU的有效推断

获取原文

摘要

Deep convolutional neural networks are well-known for the extensive parameters and computation intensity. Structured pruning is an effective solution to obtain a more compact model for the efficient inference on GPGPUs, without designing specific hardware accelerators. However, previous works resort to certain metrics in channel/filter pruning and count on labor intensive fine-tunings to recover the accuracy loss. The “inception” of the pruned model, as another form factor, has indispensable impact to the final accuracy but its importance is often ignored in these works. In this paper, we prove that optimal inception will be more likely to induce a satisfied performance and shortened fine-tuning iterations. We also propose a reinforcement learning based solution, termed as HeadStart, seeking to learn the best way of pruning aiming at the optimal inception. With the help of the specialized head-start network, it could automatically balance the tradeoff between the final accuracy and the preset speedup rather than tilting to one of them, which makes it differentiated from existing works as well. Experimental results show that HeadStart could attain up to 2. 25x inference speedup with only 1.16% accuracy loss tested with large scale images on various GPGPUs, and could be well generalized to various cutting-edge DCNN models.
机译:深度卷积神经网络以广泛的参数和计算强度众所周知。结构化修剪是一种有效的解决方案,可以获得更紧凑的GPGPUS上有效推断的模型,而无需设计特定的硬件加速器。然而,以前的作品诉诸频道/过滤器修剪的某些指标,并计算劳动密集型微调,以恢复精度损失。作为另一种形式因素的修剪模型的“成立”对最终准确性具有必不可少的影响,但在这些作品中往往忽略其重要性。在本文中,我们证明了最佳的成立将更有可能诱导满意的性能和缩短的微调迭代。我们还提出了一项基于加强学习的解决方案,称为朝证,寻求学习旨在最佳成立的最佳驯化方式。在专业的头部启动网络的帮助下,它可以自动平衡最终精度与预设加速之间的权衡而不是倾斜到其中一个,这也使其与现有工作区别不同。实验结果表明,Headstart可以获得高达2. 25x推断加速,只有1.16%的精度损耗,在各种GPGPU上测试大规模图像,并且可以很好地推广到各种尖端DCNN模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号