...
首页> 外文期刊>Neural Networks: The Official Journal of the International Neural Network Society >ADA-Tucker: Compressing deep neural networks via adaptive dimension adjustment tucker decomposition
【24h】

ADA-Tucker: Compressing deep neural networks via adaptive dimension adjustment tucker decomposition

机译:ADA-Tucker:通过自适应尺寸调整压缩深度神经网络Tucker分解

获取原文
获取原文并翻译 | 示例
           

摘要

Despite recent success of deep learning models in numerous applications, their widespread use on mobile devices is seriously impeded by storage and computational requirements. In this paper, we propose a novel network compression method called Adaptive Dimension Adjustment Tucker decomposition (ADA-Tucker). With learnable core tensors and transformation matrices, ADA-Tucker performs Tucker decomposition of arbitrary-order tensors. Furthermore, we propose that weight tensors in networks with proper order and balanced dimension are easier to be compressed. Therefore, the high flexibility in decomposition choice distinguishes ADA-Tucker from all previous low-rank models. To compress more, we further extend the model to Shared Core ADA-Tucker (SCADA-Tucker) by defining a shared core tensor for all layers. Our methods require no overhead of recording indices of non-zero elements. Without loss of accuracy, our methods reduce the storage of LeNet-5 and LeNet-300 by ratios of 691x and 233x, respectively, significantly outperforming state of the art. The effectiveness of our methods is also evaluated on other three benchmarks (CIFAR-10, SVHN, ILSVRC12) and modern newly deep networks (ResNet, Wide-ResNet). (C) 2018 Elsevier Ltd. All rights reserved.
机译:尽管近期众多应用中最近的深度学习模型成功,但他们对移动设备的广泛使用受到存储和计算要求的严重影响。在本文中,我们提出了一种新颖的网络压缩方法,称为自适应尺寸调节塔克分解(ADA-Tucker)。通过学习核心张量和转换矩阵,ADA-Tucker执行任意阶张量的Tucker分解。此外,我们提出具有适当顺序和平衡维度的网络中的重量张量更容易被压缩。因此,分解选择的高灵活性将ADA-Tucker与所有先前的低秩模型区分开来。要更换更多,我们将通过为所有层定义共享核心张量来将模型扩展为共享核心ADA-Tucker(SCADA-Tucker)。我们的方法不需要记录非零元素的记录指数的开销。不损失准确性,我们的方法分别通过691倍和233倍的比率降低LENET-5和LENET-300的储存,其明显优于现有技术。我们的方法的有效性也在其他三个基准(CIFAR-10,SVHN,ILSVRC12)和现代新深度网络(Reset,Wide-Reset)上进行评估。 (c)2018年elestvier有限公司保留所有权利。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号