首页> 外文期刊>JMLR: Workshop and Conference Proceedings >Compressing Neural Networks using the Variational Information Bottleneck
【24h】

Compressing Neural Networks using the Variational Information Bottleneck

机译:使用变分信息瓶颈压缩神经网络

获取原文
       

摘要

Neural networks can be compressed to reduce memory and computational requirements, or to increase accuracy by facilitating the use of a larger base architecture. In this paper we focus on pruning individual neurons, which can simultaneously trim model size, FLOPs, and run-time memory. To improve upon the performance of existing compression algorithms we utilize the information bottleneck principle instantiated via a tractable variational bound. Minimization of this information theoretic bound reduces the redundancy between adjacent layers by aggregating useful information into a subset of neurons that can be preserved. In contrast, the activations of disposable neurons are shut off via an attractive form of sparse regularization that emerges naturally from this framework, providing tangible advantages over traditional sparsity penalties without contributing additional tuning parameters to the energy landscape. We demonstrate state-of-the-art compression rates across an array of datasets and network architectures.
机译:可以压缩神经网络以减少内存和计算需求,或者通过促进使用更大的基础体系结构来提高准确性。在本文中,我们着重于修剪单个神经元,这些神经元可以同时修剪模型大小,FLOP和运行时内存。为了改善现有压缩算法的性能,我们利用了通过易处理的变差边界实例化的信息瓶颈原理。通过将有用的信息聚合到可以保留的神经元子集中,此信息理论界限的最小化减少了相邻层之间的冗余。相比之下,一次性的神经元的激活通过一种有吸引力的稀疏正则化形式来关闭,稀疏正则化从该框架中自然出现,提供了优于传统稀疏惩罚的明显优势,而没有为能源格局贡献额外的调整参数。我们展示了一系列数据集和网络体系结构的最新压缩率。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号