首页> 外文会议>2012 Third International Conference on Networking and Computing. >Asynchronous Memory Machine Models with Barrier Synchronization
【24h】

Asynchronous Memory Machine Models with Barrier Synchronization

机译:具有屏障同步的异步内存机器模型

获取原文
获取原文并翻译 | 示例

摘要

The Discrete Memory Machine (DMM) and the Unified Memory Machine (UMM) are theoretical parallel computing models that capture the essence of the shared memory and the global memory of GPUs. It was assumed that warps (i.e. groups of threads) on the DMM and the UMM work synchronously in the round-robin manner. However, warps work asynchronously in the actual GPUs, in the sense that warps may be randomly (or arbitrarily) dispatched for execution. The first contribution of this paper is to introduce an asynchronous version of the DMM and the UMM, in which warps are arbitrarily dispatched. Instead, we assume that threads can execute the ``sync threads'' instruction for barrier synchronization. Since the barrier synchronization operation is costly, we should evaluate and minimize the number of barrier synchronization operations performed by parallel algorithms. The second contribution of this paper is to show a parallel algorithm to compute the sum of boldmath$n$ numbers in optimal computing time and few barrier synchronization steps. Our parallel algorithm computes the sum of boldmath$n$ numbers in boldmath$O({nover w}+llog n)$ time units and boldmath$O(log{lover w}+loglog w)$ barrier synchronization steps using boldmath$wl$ threads both on the asynchronous DMM and on the asynchronous UMM with width boldmath$w$ and latency boldmath$l$. We also prove that the computing time is optimal because it matches the theoretical lower bound. Quite surprisingly, the number of barrier synchronization steps and the number of threads are independent of boldmath$n$. Even if the input size boldmath$n$ is quite large, our parallel algorithm computes the sum in optimal time units and a fixed number of sync threads using a fixed number of threads.
机译:离散内存机(DMM)和统一内存机(UMM)是理论上的并行计算模型,可捕获GPU共享内存和全局内存的本质。假定DMM和UMM上的扭曲(即线程组)以循环方式同步工作。但是,从实际意义上讲,扭曲可能是随机(或任意)调度执行的,因此在实际的GPU中异步运行。本文的第一篇贡献是介绍了DMM和UMM的异步版本,其中任意分配了扭曲。相反,我们假设线程可以执行``同步线程''指令以进行屏障同步。由于屏障同步操作的成本很高,因此我们应该评估并最小化并行算法执行的屏障同步操作的数量。本文的第二个贡献是展示了一种并行算法,可以在最佳计算时间和较少的障碍同步步骤中计算出boldmath $ n $数的总和。我们的并行算法使用boldmath $ wl计算以boldmath $ O({nover w} + llog n)$时间单位和boldmath $ O(log {lover w} + loglog w)$障碍同步步骤生成的boldmath $ n $数字和$在异步DMM和异步UMM上都具有宽度为boldmath $ w $和等待时间为boldmath $ l $的线程。我们还证明了计算时间是最佳的,因为它与理论下限匹配。令人惊讶的是,屏障同步步骤的数量和线程的数量与boldmath $ n $无关。即使输入大小boldmath $ n $很大,我们的并行算法也会以最佳时间单位和固定数量的同步线程来计算总和。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号