...
首页> 外文期刊>Parallel Processing Letters >A Novel Multi-GPU Parallel Optimization Model for The Sparse Matrix-Vector Multiplication
【24h】

A Novel Multi-GPU Parallel Optimization Model for The Sparse Matrix-Vector Multiplication

机译:稀疏矩阵-向量乘法的新型多GPU并行优化模型

获取原文
获取原文并翻译 | 示例
   

获取外文期刊封面封底 >>

       

摘要

Accelerating the sparse matrix-vector multiplication (SpMV) on the graphics processing units (GPUs) has attracted considerable attention recently. We observe that on a specific multiple-GPU platform, the SpMV performance can usually be greatly improved when a matrix is partitioned into several blocks according to a predetermined rule and each block is assigned to a GPU with an appropriate storage format. This motivates us to propose a novel multi-GPU parallel SpMV optimization model. Our model involves two stages. In the first stage, a simple rule is defined to divide any given matrix among multiple GPUs, and then a performance model, which is independent of the problems and dependent on the resources of devices, is proposed to accurately predict the execution time of SpMV kernels. Using these models, we construct in the second stage an optimally multi-GPU parallel SpMV algorithm that is automatically and rapidly generated for the platform for any problem. Given that our model for SpMV is general, independent of the problems, and dependent on the resources of devices, this model is constructed only once for each type of GPU. The experiments validate the high efficiency of our proposed model.
机译:最近,在图形处理单元(GPU)上加速稀疏矩阵矢量乘法(SpMV)引起了相当大的关注。我们观察到,在特定的多GPU平台上,将矩阵根据预定规则划分为多个块并将每个块分配给具有适当存储格式的GPU时,通常可以大大提高SpMV性能。这激励我们提出一种新颖的多GPU并行SpMV优化模型。我们的模型涉及两个阶段。在第一阶段,定义一个简单的规则以在多个GPU之间划分任何给定的矩阵,然后提出一个独立于问题并依赖于设备资源的性能模型,以准确预测SpMV内核的执行时间。使用这些模型,我们在第二阶段构造了最优的多GPU并行SpMV算法,该算法可针对任何问题自动,快速地为平台生成。鉴于我们针对SpMV的模型是通用的,与问题无关,并且取决于设备的资源,因此对于每种类型的GPU,该模型仅构建一次。实验验证了我们提出的模型的高效率。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号