首页> 外文期刊>Computer science >MVAPICH2-GPU: optimized GPU to GPU communication for InfiniBand clusters
【24h】

MVAPICH2-GPU: optimized GPU to GPU communication for InfiniBand clusters

机译:MVAPICH2-GPU:针对InfiniBand集群优化了GPU与GPU的通信

获取原文
获取原文并翻译 | 示例
       

摘要

Data parallel architectures, such as General Purpose Graphics Units (GPGPUs) have seen a tremendous rise in their application for High End Computing. However, data movement in and out of GPGPUs remain the biggest hurdle to overall performance and programmer productivity. Applications executing on a cluster with GPUs have to manage data movement using CUDA in addition to MPI, the de-facto parallel programming standard. Currently, data movement with CUDA and MPI libraries is not integrated and it is not as efficient as possible. In addition, MPI-2 one sided communication does not work for windows in GPU memory, as there is no way to remotely get or put data from GPU memory in a one-sided manner. In this paper, we propose a novel MPI design that integrates CUDA data movement transparently with MPI. The programmer is presented with one MPI interface that can communicate to and from GPUs. Data movement from GPU and network can now be overlapped. The proposed design is incorporated into the MVAPICH2 library. To the best of our knowledge, this is the first work of its kind to enable advanced MPI features and optimized pipelining in a widely used MPI library. We observe up to 45% improvement in one-way latency. In addition, we show that collective communication performance can be improved significantly: 32%, 37% and 30% improvement for Scatter, Gather and Allotall collective operations, respectively. Further, we enable MPI-2 one sided communication with GPUs. We observe up to 45% improvement for Put and Get operations.
机译:诸如通用图形单元(GPGPU)之类的数据并行体系结构在高端计算中的应用得到了极大的发展。但是,进出GPGPU的数据仍然是整体性能和程序员生产力的最大障碍。在带有GPU的集群上执行的应用程序除了必须使用MPI(事实上的并行编程标准)之外,还必须使用CUDA来管理数据移动。当前,尚未将具有CUDA和MPI库的数据移动集成在一起,并且效率不高。此外,MPI-2单面通信不适用于GPU内存中的窗口,因为无法单面远程获取或放置来自GPU内存的数据。在本文中,我们提出了一种新颖的MPI设计,该设计将CUDA数据移动与MPI透明集成。为程序员提供了一个MPI接口,该接口可以与GPU通信。来自GPU和网络的数据移动现在可以重叠。拟议的设计已合并到MVAPICH2库中。据我们所知,这是在广泛使用的MPI库中启用高级MPI功能和优化流水线的同类工作。我们观察到单向延迟最多可提高45%。此外,我们证明集体通信性能可以显着提高:Scatter,Gather和Allotall集体操作分别提高了32%,37%和30%。此外,我们启用MPI-2与GPU的单面通信。我们发现Put和Get操作最多可提高45%。

著录项

  • 来源
    《Computer science》 |2011年第4期|p.257-266|共10页
  • 作者单位

    Department of Computer Science and Engineering, The Ohio State University, Columbus, USA;

    Department of Computer Science and Engineering, The Ohio State University, Columbus, USA;

    Department of Computer Science and Engineering, The Ohio State University, Columbus, USA;

    Department of Computer Science and Engineering, The Ohio State University, Columbus, USA;

    Department of Computer Science and Engineering, The Ohio State University, Columbus, USA;

    Department of Computer Science and Engineering, The Ohio State University, Columbus, USA;

  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

    MPI; clusters; GPGPU; CUDA; infiniband;

    机译:MPI;集群GPGPU;CUDA;无限带宽;
  • 入库时间 2022-08-17 13:50:21

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号