首页> 外文会议>2011 17th IEEE International Conference on Parallel and Distributed Systems >A Static Task Scheduling Framework for Independent Tasks Accelerated Using a Shared Graphics Processing Unit
【24h】

A Static Task Scheduling Framework for Independent Tasks Accelerated Using a Shared Graphics Processing Unit

机译:使用共享图形处理单元加速独立任务的静态任务计划框架

获取原文
获取外文期刊封面目录资料

摘要

The High Performance Computing (HPC) field is witnessing the increasing use of Graphics Processing Units (GPUs) as application accelerators, due to their massively data-parallel computing architectures and exceptional floating-point computational capabilities. The performance advantage from GPU-based acceleration is primarily derived for GPU computational kernels that operate on large amount of data, consuming all of the available GPU resources. For applications that consist of several independent computational tasks that do not occupy the entire GPU, sequentially using the GPU one task at a time leads to performance inefficiencies. It is therefore important for the programmer to cluster small tasks together for sharing the GPU, however, the best performance cannot be achieved through an ad-hoc grouping and execution of these tasks. In this paper, we explore the problem of GPU tasks scheduling, to allow multiple tasks to efficiently share and be executed in parallel on the GPU. We analyze factors affecting multi-tasking parallelism and performance, followed by developing the multi-tasking execution model as a performance prediction approach. The model is validated by comparing with actual execution scenarios for GPU sharing. We then present the scheduling technique and algorithm based on the proposed model, followed by experimental verifications of the proposed approach using an NVIDIA Fermi GPU computing node. Our results demonstrate significant performance improvements using the proposed scheduling approach, compared with sequential execution of the tasks under the conventional multi-tasking execution scenario.
机译:高性能计算(HPC)领域见证了图形处理单元(GPU)作为应用加速器的越来越多的使用,这是因为它们具有大规模的数据并行计算体系结构和出色的浮点计算功能。基于GPU的加速所带来的性能优势主要来自处理大量数据,消耗所有可用GPU资源的GPU计算内核。对于包含几个不占用整个GPU的独立计算任务的应用程序,依次使用GPU一次执行一项任务会导致性能低下。因此,对于程序员来说,将小型任务聚集在一起以共享GPU非常重要,但是,通过临时分组和执行这些任务无法获得最佳性能。在本文中,我们探讨了GPU任务调度的问题,以允许多个任务有效地共享并在GPU上并行执行。我们分析影响多任务并行性和性能的因素,然后开发多任务执行模型作为性能预测方法。通过与用于GPU共享的实际执行方案进行比较来验证该模型。然后,我们基于提出的模型提出调度技术和算法,然后使用NVIDIA Fermi GPU计算节点对提出的方法进行实验验证。我们的结果表明,与传统的多任务执行方案下按顺序执行任务相比,使用建议的调度方法可以显着提高性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号