首页> 外文期刊>International journal of parallel programming >Addressing GPU On-Chip Shared Memory Bank Conflicts Using Elastic Pipeline
【24h】

Addressing GPU On-Chip Shared Memory Bank Conflicts Using Elastic Pipeline

机译:使用弹性管道解决GPU芯片上共享存储库冲突

获取原文
获取原文并翻译 | 示例

摘要

One of the major problems with the GPU on-chip shared memory is bank conflicts. We analyze that the throughput of the GPU processor core is often constrained neither by the shared memory bandwidth, nor by the shared memory latency (as long as it stays constant), but is rather due to the varied latencies caused by memory bank conflicts. This results in conflicts at the writeback stage of the in-order pipeline and causes pipeline stalls, thus degrading system throughput. Based on this observation, we investigate and propose a novel Elastic Pipeline design that minimizes the negative impact of on-chip memory bank conflicts on system throughput, by decoupling bank conflicts from pipeline stalls. Simulation results show that our proposed Elastic Pipeline together with the co-designed bank-conflict aware warp scheduling reduces the pipeline stalls by up to 64.0 % (with 42.3 % on average) and improves the overall performance by up to 20.7% (on average 13.3%) for representative benchmarks, at trivial hardware overhead.
机译:GPU片上共享内存的主要问题之一是存储体冲突。我们分析说,GPU处理器核心的吞吐量通常不受共享内存带宽或共享内存延迟(只要保持不变)的限制,而是由于内存组冲突导致的延迟不同而造成的。这将导致有序管道的写回阶段发生冲突,并导致管道停顿,从而降低系统吞吐量。基于此观察,我们研究并提出了一种新颖的弹性流水线设计,该设计通过将流水线冲突与流水线停顿脱钩,从而将片上内存条冲突对系统吞吐量的负面影响降至最低。仿真结果表明,我们提出的弹性管道与共同设计的具有银行冲突意识的翘曲调度一起,可将管道停顿减少多达64.0%(平均为42.3%),并将整体性能提高高达20.7%(平均为13.3) %)用于代表性基准测试,而琐碎的硬件开销。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号