首页> 外文会议>International Symposium on Computer Architecture and High Performance Computing >Designing High Performance Heterogeneous Broadcast for Streaming Applications on GPU Clusters
【24h】

Designing High Performance Heterogeneous Broadcast for Streaming Applications on GPU Clusters

机译:设计高性能异构广播,用于GPU集群上的流媒体应用

获取原文

摘要

High-performance streaming applications are beginning to leverage the compute power offered by graphics processing units (GPUs) and high network throughput offered by high performance interconnects such as InfiniBand (IB) to boost their performance and scalability. These applications rely heavily on broadcast operations to move data, which is stored in the host memory, from a single source-typically live-to multiple GPU-based computing sites. While homogeneous broadcast designs take advantage of IB hardware multicast feature to boost their performance, their heterogeneous counterpart requires an explicit data movement between Host and GPU, which significantly hurts the overall performance. There is a dearth of efficient heterogeneous broadcast designs for streaming applications especially on emerging multi-GPU configurations. In this work, we propose novel techniques to fully and conjointly take advantage of NVIDIA GPUDirect RDMA (GDR), CUDA inter-process communication (IPC) and IB hardware multicast features to design high-performance heterogeneous broadcast operations for modern multi-GPU systems. We propose intra-node, topology-aware schemes to maximize the performance benefits while minimizing the utilization of valuable PCIe resources. Further, we optimize the communication pipeline by overlapping the GDR + IB hardware multicast operations with CUDA IPC operations. Compared to existing solutions, our designs show up to 3X improvement in the latency of a heterogeneous broadcast operation. Our designs also show up to 67% improvement in execution time of a streaming benchmark on a GPU-dense Cray CS-Storm system with 88 GPUs.
机译:高性能流媒体应用正在开始利用由图形处理单元(GPU)提供的计算能力和网络的高吞吐量的高性能互连提供如InfiniBand(IB),以提高它们的性能和可扩展性。这些应用在很大程度上依赖于广播业务来移动数据,存储在主机内存,从单一来源 - 通常活到多个基于GPU计算的网站。虽然同质广播设计将IB硬件组播功能的优势,以提高他们的表现,他们的异构对口需要主机和GPU之间的明确数据移动,该显著伤害了整体性能。有有效的异构广播设计的用于流应用的尤其是对新出现的多GPU配置缺乏。在这项工作中,我们提出了新的技术,以充分和共同地采取NVIDIA GPUDirect RDMA(GDR)的优势,CUDA进程间通信(IPC)和IB硬件组播功能,以设计出高性能异构广播业务为现代的多GPU系统。我们建议节点内,拓扑感知方案,最大限度的性能优势,同时尽量减少宝贵的PCIe资源利用率。此外,我们通过用CUDA IPC操作重叠GDR + IB硬件组播业务优化通信管道。相比于现有的解决方案,我们的设计出现在一个异构的广播操作的等待时间提高了3倍。我们的设计还出现在一个GPU密集克雷CS-风暴系统88分的GPU上的流基准的执行时间,67%的改善。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号