首页> 外文会议>IEEE International Conference on Multimedia and Expo >Lightweight Compression Of Neural Network Feature Tensors For Collaborative Intelligence
【24h】

Lightweight Compression Of Neural Network Feature Tensors For Collaborative Intelligence

机译:轻量级压缩神经网络特征张量以实现协作智能

获取原文

摘要

In collaborative intelligence applications, part of a deep neural network (DNN) is deployed on a relatively low-complexity device such as a mobile phone or edge device, and the remainder of the DNN is processed where more computing resources are available, such as in the cloud. This paper presents a novel lightweight compression technique designed specifically to code the activations of a split DNN layer, while having a low complexity suitable for edge devices and not requiring any retraining. We also present a modified entropy-constrained quantizer design algorithm optimized for clipped activations. When applied to popular object-detection and classification DNNs, we were able to compress the 32-bit floating point activations down to 0.6 to 0.8 bits, while keeping the loss in accuracy to less than 1%. When compared to HEVC, we found that the lightweight codec consistently provided better inference accuracy, by up to 1.3%. The performance and simplicity of this lightweight compression technique makes it an attractive option for coding a layer’s activations in split neural networks for edge/cloud applications.
机译:在协作智能应用程序中,深度神经网络(DNN)的一部分部署在相对较低复杂度的设备(例如手机或边缘设备)上,其余DNN在有更多计算资源可用的情况下进行处理,例如云端。本文提出了一种新颖的轻量级压缩技术,该技术专门设计用于对拆分的DNN层的激活进行编码,同时具有适用于边缘设备且无需任何重新训练的低复杂度。我们还提出了针对限幅激活优化的改进的熵约束量化器设计算法。当应用于流行的目标检测和分类DNN时,我们能够将32位浮点激活压缩到0.6至0.8位,同时将精度损失保持在1%以内。与HEVC相比,我们发现轻量级编解码器始终提供更高的推断精度,最高可达1.3%。这种轻量级压缩技术的性能和简便性使其成为在边缘/云应用的分离式神经网络中为层的激活进行编码的有吸引力的选择。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号