首页> 外文会议>Annual IEEE/ACM International Symposium on Microarchitecture >vDNN: Virtualized deep neural networks for scalable, memory-efficient neural network design
【24h】

vDNN: Virtualized deep neural networks for scalable, memory-efficient neural network design

机译:vDNN:虚拟化深度神经网络,可扩展,内存高效的神经网络设计

获取原文

摘要

The most widely used machine learning frameworks require users to carefully tune their memory usage so that the deep neural network (DNN) fits into the DRAM capacity of a GPU. This restriction hampers a researcher's flexibility to study different machine learning algorithms, forcing them to either use a less desirable network architecture or parallelize the processing across multiple GPUs. We propose a runtime memory manager that virtualizes the memory usage of DNNs such that both GPU and CPU memory can simultaneously be utilized for training larger DNNs. Our virtualized DNN (vDNN) reduces the average GPU memory usage of AlexNet by up to 89%, OverFeat by 91%, and GoogLeNet by 95%, a significant reduction in memory requirements of DNNs. Similar experiments on VGG-16, one of the deepest and memory hungry DNNs to date, demonstrate the memory-efficiency of our proposal. vDNN enables VGG-16 with batch size 256 (requiring 28 GB of memory) to be trained on a single NVIDIA Titan X GPU card containing 12 GB of memory, with 18% performance loss compared to a hypothetical, oracular GPU with enough memory to hold the entire DNN.
机译:最广泛使用的机器学习框架要求用户仔细调整其内存使用情况,以便深度神经网络(DNN)可以适合GPU的DRAM容量。这种限制妨碍了研究人员研究不同机器学习算法的灵活性,从而迫使他们要么使用不太理想的网络体系结构,要么跨多个GPU并行处理。我们提出了一个运行时内存管理器,用于虚拟化DNN的内存使用情况,以便GPU和CPU内存可以同时用于训练更大的DNN。我们的虚拟化DNN(vDNN)将AlexNet的平均GPU内存使用量减少了89%,OverFeat减少了91%,GoogLeNet减少了95%,从而大大降低了DNN的内存需求。在VGG-16上进行的类似实验(迄今为止最深入的内存不足DNN)之一,证明了我们建议的存储效率。 vDNN使批量大小为256(需要28 GB内存)的VGG-16可以在包含12 GB内存的单块NVIDIA Titan X GPU卡上进行训练,与假设的,具有足够内存以容纳的假眼式GPU相比,性能损失为18%整个DNN。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号