首页> 外文会议>International Conference on Human Computer Interactions >Multi-Model Inference Acceleration on Embedded Multi-Core Processors
【24h】

Multi-Model Inference Acceleration on Embedded Multi-Core Processors

机译:嵌入式多核处理器上的多模型推理加速

获取原文

摘要

The predominant resource efficient approaches that enable on-device inference include designing lightweight DNN architectures like MobileNets, SqueezeNets, compressing model using techniques such as network pruning, vector quantization, distillation, binarization. Recent research on using dynamic layer-wise partitioning and partial execution of CNN based model inference also make it possible to co-inference on memory and computation resource constrained devices. However, these approaches have their own bottleneck, lightweight DNN architectures and model compression usually compromise accuracy in order to deploy on resource constrained devices, dynamic model partitioning efficiency depends heavily on the network condition. This paper proposes an approach for multimodel inference acceleration on heterogeneous devices. The idea is to deploy multiple single object detection model instead of one heavy multiple object, this is because in most cases it only needs to detect one or two objects in one scenario and single object detection model weight could be lighter for the same resolution quality and require less resource. Moreover, in cloud-edge-device scenario, with the help of a scheduler policy, it is possible to gradually update models in need.
机译:使能On-Device推断的主要资源有效方法包括设计MobileNets,SheeezEnets,压缩模型等轻量级DNN架构,使用诸如网络修剪,矢量量化,蒸馏,二值化等技术。最近使用基于CNN的模型推断的动态层明智的分区和部分执行的研究还使得可以在存储器和计算资源受限设备上共同推断。但是,这些方法具有自己的瓶颈,轻量级DNN架构和模型压缩通常会损害精度,以便在资源受限设备上部署,动态模型分区效率在很大程度上取决于网络条件。本文提出了一种对异构装置的多模型推理加速的方法。这个想法是部署多个单个对象检测模型而不是一个沉重的多个对象,这是因为在大多数情况下,它只需要在一个场景中检测一个或两个对象,并且单个对象检测模型重量对于相同的分辨率质量和单个对象检测模型重量可能会更轻。需要更少的资源。此外,在云边缘设备方案中,在调度程序策略的帮助下,可以逐步更新需要的模型。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号