...
首页> 外文期刊>Computers & Fluids >High order accurate simulation of compressible flows on GPU clusters over Software Distributed Shared Memory
【24h】

High order accurate simulation of compressible flows on GPU clusters over Software Distributed Shared Memory

机译:通过软件分布式共享内存对GPU群集上的可压缩流进行高阶准确仿真

获取原文
获取原文并翻译 | 示例

摘要

The advent of multicore processors during the past decade and especially the recent introduction of many-core Graphics Processing Units (GPUs) open new horizons to large-scale, high-resolution simulations for a broad range of scientific fields. Residing at the forefront of advancements in multiprocessor technology, GPUs are often chosen as co-processors when intensive parts of applications need to be computed. Among the various domains, the scientific area of Computational Fluid Dynamics (CFD) is a potential candidate that could significantly benefit from the utilization of many-core GPUs. In order to investigate this possibility, we herein evaluate the performance of a high order accurate method for the simulation of compressible flows. Targeting computer systems with multiple GPUs, the current implementation and the respective performance evaluation are taking place on a GPU cluster. With respect to using these GPUs, this paper offers an alternative to the mainstream approach of message passing by considering shared memory abstraction. In the implementations presented in this paper, the updates on shared data are not explicitly coded by the programmer across the simulation phases, but are propagated through Software Distributed Shared Memory (SDSM). This way, we intend to preserve a unified memory view that extends the memory hierarchy from the node level to the cluster level. Such an extension could significantly facilitate the porting of multithreaded codes at GPU clusters. Our results indicate that the presented approach is competitive with the message passing paradigm and they lay grounds for further research on the use of shared memory abstraction for future GPU clusters.
机译:在过去的十年中,多核处理器的出现,尤其是最近推出的多核图形处理单元(GPU),为广泛的科学领域的大规模,高分辨率仿真打开了新的视野。站在多处理器技术发展的最前沿,当需要计算大量应用程序时,通常会选择GPU作为协处理器。在各个领域中,计算流体力学(CFD)的科学领域是一个潜在的候选者,可以从多核GPU的使用中显着受益。为了研究这种可能性,我们在这里评估用于可压缩流模拟的高阶精确方法的性能。针对具有多个GPU的计算机系统,当前的实现以及相应的性能评估都在GPU集群上进行。关于使用这些GPU,本文通过考虑共享内存抽象为主流的消息传递方法提供了一种替代方法。在本文介绍的实现中,程序员在模拟阶段未对共享数据的更新进行明确编码,而是通过软件分布式共享内存(SDSM)进行传播。这样,我们打算保留一个统一的内存视图,该视图将内存层次结构从节点级别扩展到群集级别。这种扩展可以显着促进在GPU群集上移植多线程代码。我们的结果表明,所提出的方法与消息传递范例具有竞争性,它们为在将来的GPU群集中使用共享内存抽象的进一步研究奠定了基础。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号