首页> 外文会议>International Conference on Computational Performance Evaluation >Techniques for Compressing Deep Convolutional Neural Network
【24h】

Techniques for Compressing Deep Convolutional Neural Network

机译:深度卷积神经网络的压缩技术

获取原文

摘要

Deep Convolutional Neural Network (CNNs) has evolved to be a progressive technique for both artificial intelligence and computer vision applications. However, these models are also known for being large, computationally expensive and requires lot of memory therefore its hard to implement them on various embedded systems or mobile phones as they have limited hardware resources, low power budget and strict latency requirements. This has motivated researchers to develop various techniques to achieve the goal of model compression of deep CNNs and also to simultaneously optimize its performance. In this paper, recent bench marking advancements in the techniques of pruning, quantization and coding are systematically explored which are mostly implemented on the top of convolutional neural networks with convolutional layer and fully connected layer to provide a compressed network. Also the challenges and future directions are discussed to reduce the model size.
机译:深度卷积神经网络(CNN)已发展成为一种针对人工智能和计算机视觉应用程序的先进技术。但是,这些模型也因其庞大,计算量大且需要大量内存而闻名,因此,由于它们具有有限的硬件资源,低功耗预算和严格的延迟要求,因此很难在各种嵌入式系统或移动电话上实现它们。这激励了研究人员开发各种技术,以实现对深层CNN进行模型压缩的目标,并同时优化其性能。在本文中,系统地研究了修剪,量化和编码技术中最新的基准标记技术进步,这些技术主要在具有卷积层和完全连接层的卷积神经网络的顶部实现,以提供压缩网络。还讨论了减少模型尺寸的挑战和未来方向。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号