首页> 外文会议>International conference on web information systems and applications >Inference Acceleration Model of Branched Neural Network Based on Distributed Deployment in Fog Computing
【24h】

Inference Acceleration Model of Branched Neural Network Based on Distributed Deployment in Fog Computing

机译:基于分布式部署的分支神经网络推理加速模型在雾计算中的分布式部署

获取原文

摘要

Research based on deep neural networks (DNN) is becoming more common. In order to solve the problem that DNN needs to consume a lot of performance during the use prediction process and generate unacceptable delays for users, a distributed neural network deployment model based on fog computing is proposed. The distributed deployment of deep neural networks in fog computing scenarios is analyzed. A deployment algorithm based on Solution Space Tree Pruning (SSTP) is designed, and a suitable fog computing node deployment model is selected to reduce the delay of prediction tasks. An algorithm for Maximizing Accuracy based on Guaranteed Latency (MAL) is designed and implemented, and suitable fog computing nodes are selected for different tasks to exit the prediction task. Simulation experiment results show that compared with the method of deploying neural network models in the cloud, the model prediction delay of the distributed neural network model based on fog computing is reduced by an average of 44.79%. Reduced the average computing acceleration framework of similar algorithms by 28.75%.
机译:基于深度神经网络(DNN)的研究变得越来越普遍。为了解决DNN在使用预测过程中需要消耗大量性能并为用户生成不可接受的延迟而产生的问题,提出了一种基于雾计算的分布式神经网络部署模型。分析了雾计算场景中深神经网络的分布式部署。设计了一种基于解决方案空间树修剪(SSTP)的部署算法,选择合适的雾计算节点部署模型以减少预测任务的延迟。设计和实现了一种基于保证延迟(MAL)最大化精度的算法,为不同任务选择合适的雾计算节点以退出预测任务。仿真实验结果表明,与在云中部署神经网络模型的方法相比,基于雾计算的分布式神经网络模型的模型预测延迟平均降低了44.79%。将相似算法的平均计算加速框架降低28.75%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号