首页> 外文期刊>高技术通讯(英文版) >Optimizing deep learning inference on mobile devices with neural network accelerators
【24h】

Optimizing deep learning inference on mobile devices with neural network accelerators

机译:使用神经网络加速器优化移动设备上的深度学习推理

获取原文
获取原文并翻译 | 示例
       

摘要

Deep learning has now been widely used in intelligent apps of mobile devices.In pursuit of ultra-low power and latency,integrating neural network accelerators(NNA)to mobile phones has become a trend.However,conventional deep learning programming frameworks are not well-developed to support such devices,leading to low computing efficiency and high memory-occupation.To address this problem,a 2-stage pipeline is proposed for optimizing deep learning model inference on mobile devices with NNAs in terms of both speed and memory-footprint.The 1 st stage reduces computation workload via graph optimization,including splitting and merging nodes.The 2 nd stage goes further by optimizing at compilation level,including kernel fusion and in-advance compilation.The proposed optimizations on a commercial mobile phone with an NNA is evaluated.The experimental results show that the proposed approaches achieve 2.8×to 26×speed up,and reduce the memory-footprint by up to 75%.

著录项

  • 来源
    《高技术通讯(英文版)》 |2019年第4期|417-425|共9页
  • 作者

    Zeng Xi; Xu Yunlong; Zhi Tian;

  • 作者单位

    Intelligent Processor Research Center Institute of Computing Technology Chinese Academy of Sciences Beijing 100190 P.R.China;

    University of Chinese Academy of Sciences Beijing 100049 P.R.China;

    Cambricon Technologies Corporation Limited Beijing 100191 P.R.China;

  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

  • 入库时间 2022-08-19 04:31:36
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号