【24h】

Scalable and Fast Lazy Persistency on GPUs

机译:GPU上的可扩展和快速懒惰持久性

获取原文

摘要

GPUs applications, including many scientific and machine learning applications, increasingly demand larger memory capacity. NVM is promising higher density compared to DRAM and better future scaling potentials. Long running GPU applications can benefit from NVM by exploiting its persistency, allowing crash recovery of data in memory. In this paper, we propose mapping Lazy Persistency (LP) to GPUs and identify the design space of such mapping. We then characterize LP performance on GPUs, varying the checksum type, reduction method, use of locking, and hash table designs. Armed with insights into the performance bottlenecks, we propose a hash table-less method that performs well on hundreds and thousands of threads, achieving persistency with nearly negligible (2.1%) slowdown for a variety of representative benchmarks. We also propose a directive-based programming language support to simplify programming effort for adding LP to GPU applications.
机译:GPU应用程序(包括许多科学和机器学习应用程序)越来越需要更大的内存容量。与DRAM相比,NVM有望实现更高的密度并具有更好的未来扩展潜力。长时间运行的GPU应用程序可以通过利用NVM的持久性来从NVM中受益,从而可以对内存中的数据进行崩溃恢复。在本文中,我们建议将懒惰持久性(LP)映射到GPU,并确定这种映射的设计空间。然后,我们表征GPU上的LP性能,改变校验和类型,减少方法,使用锁定和哈希表设计。借助对性能瓶颈的洞察力,我们提出了一种无哈希表的方法,该方法在成千上万的线程上表现良好,对于各种代表性基准测试,其持久性几乎可以忽略不计(2.1%)。我们还提出了一种基于指令的编程语言支持,以简化将LP添加到GPU应用程序的编程工作。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号