首页> 外文期刊>ETRI journal >Deep compression of convolutional neural networks with low‐rank approximation
【24h】

Deep compression of convolutional neural networks with low‐rank approximation

机译:低秩逼近的卷积神经网络深度压缩

获取原文
           

摘要

The application of deep neural networks (DNNs) to connect the world with cyber physical systems (CPSs) has attracted much attention. However, DNNs require a large amount of memory and computational cost, which hinders their use in the relatively low‐end smart devices that are widely used in CPSs. In this paper, we aim to determine whether DNNs can be efficiently deployed and operated in low‐end smart devices. To do this, we develop a method to reduce the memory requirement of DNNs and increase the inference speed, while maintaining the performance (for example, accuracy) close to the original level. The parameters of DNNs are decomposed using a hybrid of canonical polyadic–singular value decomposition, approximated using a tensor power method, and fine‐tuned by performing iterative one‐shot hybrid fine‐tuning to recover from a decreased accuracy. In this study, we evaluate our method on frequently used networks. We also present results from extensive experiments on the effects of several fine‐tuning methods, the importance of iterative fine‐tuning, and decomposition techniques. We demonstrate the effectiveness of the proposed method by deploying compressed networks in smartphones.
机译:深度神经网络(DNN)将世界与网络物理系统(CPS)连接的应用备受关注。但是,DNN需要大量的内存和计算成本,这阻碍了它们在CPS中广泛使用的相对低端的智能设备中的使用。本文旨在确定DNN是否可以在低端智能设备中有效部署和运行。为此,我们开发了一种方法来减少DNN的内存需求并提高推理速度,同时保持性能(例如准确性)接近原始水平。 DNN的参数使用规范的多-奇异值分解的混合分解,使用张量幂方法进行近似,并通过执行迭代单次混合细调进行微调以从降低的精度中恢复。在这项研究中,我们评估了常用网络上的方法。我们还提供了关于几种微调方法的效果,迭代微调的重要性以及分解技术的广泛实验的结果。通过在智能手机中部署压缩网络,我们证明了该方法的有效性。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号