首页> 外文会议>IEEE International Conference on Big Data >Towards Power Efficiency in Deep Learning on Data Center Hardware
【24h】

Towards Power Efficiency in Deep Learning on Data Center Hardware

机译:在数据中心硬件上进行深度学习以提高电源效率

获取原文

摘要

Deep learning (DL) is a computationally intensive workload that is expected to grow rapidly in data centers in the near future. Its high energy demand necessitates finding ways to improve computational efficiency. In this work, we directly measure power used by the whole system as well as that used by GPU, CPU, and RAM during DL training to determine their contributions to the overall energy consumption. We find that while GPUs use most of the power – about 70 % - the consumption of other components is also significant and their optimizations can bring important power savings. Evaluating a multitude of options, we identify the parameters that bring in the most power savings. Overall, an energy savings of over 20% of can be obtained by adjusting system settings alone without changing the workload, at the cost of a minor increase in runtime. Alternatively, if runtime needs to stay constant, an 18% energy savings is identified. In distributed multi-server DL, we find that scale-out overhead has only a small energy cost, making distributed training more energy-efficient than expected. Implications for the field and ways to make DL more energy-efficient going forward are also discussed. (Abstract)
机译:深度学习(DL)是计算密集型工作负载,预计在不久的将来会在数据中心迅速增长。它对能源的需求很高,因此必须找到提高计算效率的方法。在这项工作中,我们直接测量整个系统以及DL训练期间GPU,CPU和RAM所使用的功率,以确定它们对总体能耗的贡献。我们发现,尽管GPU消耗了大部分功率(约70%),但其他组件的消耗也很可观,其优化可以节省大量功率。通过评估众多选项,我们确定了可最大程度节省功率的参数。总体而言,仅通过调整系统设置而不改变工作量即可节省超过20%的能源,但代价是运行时间会略有增加。另外,如果运行时间需要保持恒定,则可以节省18%的能源。在分布式多服务器DL中,我们发现横向扩展开销只有很少的能源成本,这使得分布式培训比预期的能效更高。还讨论了对该领域的影响以及使DL更加节能的方法。 (抽象的)

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号