首页> 外文会议>IEEE Computer Society Annual Symposium on VLSI >Reducing the Model Order of Deep Neural Networks Using Information Theory
【24h】

Reducing the Model Order of Deep Neural Networks Using Information Theory

机译:利用信息论降低深层神经网络的模型阶数

获取原文

摘要

Deep neural networks are typically represented by a much larger number of parameters than shallow models, making them prohibitive for small footprint devices. Recent research shows that there is considerable redundancy in the parameter space of deep neural networks. In this paper, we propose a method to compress deep neural networks by using the Fisher Information metric, which we estimate through a stochastic optimization method that keeps track of second-order information in the network. We first remove unimportant parameters and then use non-uniform fixed point quantization to assign more bits to parameters with higher Fisher Information estimates. We evaluate our method on a classification task with a convolutional neural network trained on the MNIST data set. Experimental results show that our method outperforms existing methods for both network pruning and quantization.
机译:深度神经网络通常由比浅模型更大数量的参数表示,使得它们对小型足迹装置禁止。最近的研究表明,深神经网络的参数空间中存在相当大的冗余。在本文中,我们提出了一种通过使用Fisher信息度量来压缩深神经网络的方法,我们通过随机优化方法估算,该方法通过跟踪网络中的二阶信息。我们首先删除不重要的参数,然后使用不均匀的固定点量化来将更多位分配给具有更高Fisher信息估算的参数。我们在具有在Mnist数据集上培训的卷积神经网络的分类任务中评估我们的方法。实验结果表明,我们的方法优于网络修剪和量化的现有方法。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号