首页> 美国卫生研究院文献>Sensors (Basel Switzerland) >Implementation of Lightweight Convolutional Neural Networks via Layer-Wise Differentiable Compression
【2h】

Implementation of Lightweight Convolutional Neural Networks via Layer-Wise Differentiable Compression

机译:通过三层可微分压缩实现轻质卷积神经网络

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

Convolutional neural networks (CNNs) have achieved significant breakthroughs in various domains, such as natural language processing (NLP), and computer vision. However, performance improvement is often accompanied by large model size and computation costs, which make it not suitable for resource-constrained devices. Consequently, there is an urgent need to compress CNNs, so as to reduce model size and computation costs. This paper proposes a layer-wise differentiable compression (LWDC) algorithm for compressing CNNs structurally. A differentiable selection operator OS is embedded in the model to compress and train the model simultaneously by gradient descent in one go. Instead of pruning parameters from redundant operators by contrast to most of the existing methods, our method replaces the original bulky operators with more lightweight ones directly, which only needs to specify the set of lightweight operators and the regularization factor in advance, rather than the compression rate for each layer. The compressed model produced by our method is generic and does not need any special hardware/software support. Experimental results on CIFAR-10, CIFAR-100 and ImageNet have demonstrated the effectiveness of our method. LWDC obtains more significant compression than state-of-the-art methods in most cases, while having lower performance degradation. The impact of lightweight operators and regularization factor on the compression rate and accuracy also is evaluated.
机译:卷积神经网络(CNNS)在各个域中取得了重大突破,例如自然语言处理(NLP)和计算机视觉。但是,性能改进通常伴随着大型模型大小和计算成本,这使得它不适合资源受限的设备。因此,迫切需要压缩CNN,以降低模型大小和计算成本。本文提出了一种用于在结构上压缩CNN的层面可分解压缩(LWDC)算法。在模型中嵌入了一个可差的选择操作OS,以便通过梯度下降来压缩和培训模型。与大多数现有方法相比,我们的方法直接用更轻质的操作员取代原始笨重运营商,而不是预先指定一组轻量级运算符和预先规范,而不是压缩每层的速率。我们的方法生产的压缩模型是通用的,不需要任何特殊的硬件/软件支持。 CiFar-10,CiFar-100和Imagenet的实验结果表明了我们方法的有效性。在大多数情况下,LWDC在最先进的方法中获得比最先进的方法更高的压缩,同时具有较低的性能下降。还评估了轻量级运营商和正则化因子对压缩率和准确性的影响。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号