首页> 外文会议>International Conference on High Performance Computing Simulation >Sparse matrix computations on clusters with GPGPUs
【24h】

Sparse matrix computations on clusters with GPGPUs

机译:与GPGPU的群集上的稀疏矩阵计算

获取原文

摘要

Hybrid nodes containing GPUs are rapidly becoming the norm in parallel machines. We have conducted some experiments regarding how to plug GPU-enabled computational kernels into PSBLAS, a MPI-based library specifically geared towards sparse matrix computations. In this paper, we present our findings on which strategies are more promising in the quest for the optimal compromise among raw performance, speedup, software maintainability, and extensibility. We consider several solutions to implement the data exchange with the GPU focusing on the data access and transfer, and present an experimental evaluation for a cluster system with up to two GPUs per node. In particular, we compare the pinned memory and the Open-MPI approaches, which are the two most used alternatives for multi-GPU communication in a cluster environment. We find that OpenMPI turns out to be the best solution for large data transfers, while the pinned memory approach is still a good solution for small transfers between GPUs.
机译:包含GPU的混合节点正在并行机器中的标准。我们对如何将支持GPU的计算内核插入PSBLA的一些实验,基于MPI的库专门针对稀疏矩阵计算。在本文中,我们在寻求原始性能,加速,软件可维护性和可扩展性方面追求最佳折衷的策略在哪些策略上展示了我们的调查结果。我们考虑使用专注于数​​据访问和传输的GPU实现数据交换的几个解决方案,并为每个节点的GPU提供了一个群集系统的实验评估。特别是,我们比较固定的内存和开放式MPI方法,这些方法是群集环境中多GPU通信的两个最使用的替代方案。我们发现OpenMPI证明是大数据传输的最佳解决方案,而固定的内存方法仍然是GPU之间的小转移的好解决方案。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号