首页> 外文会议>International Conference on Computer Supported Cooperative Work in Design >PipePar: A Pipelined Hybrid Parallel Approach for Accelerating Distributed DNN Training
【24h】

PipePar: A Pipelined Hybrid Parallel Approach for Accelerating Distributed DNN Training

机译:pipepar:一种用于加速分布式DNN训练的流水线混合并行方法

获取原文

摘要

Large scale DNN training tasks are exceedingly compute-intensive and time-consuming, which are usually executed on highly-parallel platforms. Data and model parallelization is a common way to speed up the training progress across devices. However, they tend to achieve sub-optimal performance due to the communication overheads and unbalanced load among servers. Recent emerging pipelining solutions mitigate the above issues, incorporating the advantages of data and model parallelism. In this paper, we make a step further towards optimizing the execution of pipelining. We introduce PipePar, a pipeline-parallel DNN training method that provides optimized execution strategies of layer-stacked DNNs. PipePar considers the entire tensor partition space of pipelining and explores potential hybrid parallel configurations of each stage in the pipeline. Additionally, we notice the network heterogeneity between different GPU servers and it is inevitable to transfer tensors with different bandwidths and latency. So, taking into account both computation and communication capacity of different GPU servers, PipePar is intended to find a elastic load distribution strategy at different levels. We evaluate PipePar with a set of real-world DNNs on 4 GPU servers. Our experimental results show that PipePar is able to find an efficient strategy that are up to $2.16imes$ faster than state-of-the-art hybrid parallelization approaches.
机译:大规模的DNN培训任务非常规格 - 密集和耗时,通常在高度平行的平台上执行。数据和模型并行化是加快设备培训培训进度的常用方式。然而,由于服务器之间的通信开销和不平衡负载,它们倾向于实现次优性能。最近的新兴流水线解决方案减轻了上述问题,包括数据和模型并行性的优点。在本文中,我们进一步迈向优化流水线的执行。我们介绍Pipepar,一种管道平行DNN训练方法,提供了层堆叠DNN的优化执行策略。 pipepar认为流水线的整个张量分区空间,并探讨了管道中每个阶段的潜在混合并行配置。此外,我们注意到不同GPU服务器之间的网络异质性,并且不可避免地传输具有不同带宽和延迟的张量。因此,考虑到不同GPU服务器的计算和通信能力,PIPE1p旨在在不同级别找到弹性负载分布策略。我们在4个GPU服务器上评估带有一组真实DNN的PIPEAP1。我们的实验结果表明,Pipepar能够找到一个有效的策略 $ 2.16 times $ 比最先进的混合并行化方法更快。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号