首页> 外文会议> >A Neural Network Model for Cache and Memory Prediction of Neural Networks
【24h】

A Neural Network Model for Cache and Memory Prediction of Neural Networks

机译:用于神经网络缓存和内存预测的神经网络模型

获取原文
获取原文并翻译 | 示例

摘要

Neural networks have been widely applied to various research and production fields. However, most recent research is focused on the establishment and selection of a specific neural network model. Less attention is paid to their system overhead despite of their massive computing and storage resource demand. This research focuses on a relatively new research direction that models the system-level memory and cache demand of neural networks. We utilize a neural network to learn and predict hit ratio curve and memory footprint of neural networks with their hyper-parameters as input. The prediction result is used to drive cache partitioning and memory partitioning to optimize co-execution of multiple neural networks. To demonstrate effectiveness of our approach, we model four common networks, BP neural network, convolutional neural network, recurrent neural network, and autoencoder. We investigate the influence of hyper-parameters of each model on the last level cache and memory demand. We resort to the BP algorithm as the learning tool to predict last level cache hit ratio curve and memory usage. Our experimental results show that cache and memory allocation schemes guided by our prediction optimize for a wide range of performance targets.
机译:神经网络已广泛应用于各种研究和生产领域。但是,最近的研究集中在特定神经网络模型的建立和选择上。尽管对计算和存储资源的需求量很大,但对系统开销的关注却较少。这项研究集中在一个相对较新的研究方向,该方向对神经网络的系统级内存和缓存需求进行建模。我们利用神经网络以其超参数作为输入来学习和预测命中率曲线和神经网络的内存占用量。预测结果用于驱动缓存分区和内存分区,以优化多个神经网络的共执行。为了证明我们方法的有效性,我们对四个常见网络进行了建模,分别是BP神经网络,卷积神经网络,递归神经网络和自动编码器。我们调查了每个模型的超参数对最后一级缓存和内存需求的影响。我们将BP算法作为学习工具来预测上一级缓存命中率曲线和内存使用情况。我们的实验结果表明,以我们的预测为指导的缓存和内存分配方案可针对各种性能目标进行优化。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号