首页> 外文会议>Symposium on Application Accelerators in High Performance Computing >Performance of Parallel Sparse Matrix-Vector Multiplications in Linear Solves on Multiple GPUs
【24h】

Performance of Parallel Sparse Matrix-Vector Multiplications in Linear Solves on Multiple GPUs

机译:多个GPU上线性求解中并行稀疏矩阵 - 矢量乘法的性能

获取原文

摘要

Modern numerical simulations often require solving extremely large sparse linear systems. Solving these linear systems using Krylov iterative methods requires repeated sparse matrix-vector multiplications which can be the most computationally expensive part of the simulation. Since Graphics Processing Units (GPUs) provide a significant increase in floating point operations per second and memory bandwidth over conventional Central Processing Units (CPUs), performing sparse matrix-vector multiplications with these co-processors can decrease the amount of time required to solve a given linear system. In this paper, we investigate the performance of sparse matrix-vector multiplications across multiple GPUs. This is performed in the context of the solution of symmetric positive-definite linear systems using a conjugate-gradient iteration preconditioned with a least-squares polynomial preconditioner using the PETSc library.
机译:现代数值模拟通常需要解决极大的稀疏线性系统。使用Krylov迭代方法求解这些线性系统需要重复的稀疏矩阵 - 矢量乘法,这可以是模拟最昂贵的部分。由于图形处理单元(GPU)在传统的中央处理单元(CPU)上提供浮点操作和内存带宽的浮点操作显着增加,并且利用这些协处理器执行稀疏矩阵矢量乘法可以降低求解A所需的时间量给定线性系统。在本文中,我们调查多个GPU跨越稀疏矩阵矢量乘法的性能。这在使用PETSC库的使用最小二乘多项式前提者的缀合物梯度迭代的对称正定线性系统的解决方案中执行。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号