首页> 外文会议>International Conference on Parallel Processing >Declarative Tuning for Locality in Parallel Programs
【24h】

Declarative Tuning for Locality in Parallel Programs

机译:并行程序中的局部性的声明式调整

获取原文

摘要

Optimized placement of data and computation for locality is critical for improving performance and reducing energy consumption on modern computing systems. However, for most programming models, modifying data and computation placements typically requires rewriting large portions of the application, thereby posing a huge performance portability challenge in today's rapidly evolving architecture landscape. In this paper we present TunedCnC, a novel, declarative and flexible CnC tuning framework for controlling the spatial and temporal placement of data and computation by specifying hierarchical affinity groups and distribution functions. TunedCnC emphasizes a separation of concerns: the domain expert specifies a parallel application by defining data and control dependences, while the tuning expert specifies how the application should be executed on a given architecture - defining when and where for data and computation placement. The application remains unchanged when tuned for a different platform or towards different performance goals. We evaluate the utility of TunedCnC on several applications, and demonstrate that varying the tuning specification can have a significant impact on an application's performance. Our evaluation is performed using an implementation of the Concurrent Collections (CnC) declarative parallel programming model, but our results should be applicable to tuning of other data-flow task-parallel programming models as well.
机译:在本地优化数据和计算的位置对于提高性能和减少现代计算系统上的能耗至关重要。但是,对于大多数编程模型而言,修改数据和计算位置通常需要重写应用程序的大部分内容,从而在当今快速发展的架构环境中提出了巨大的性能可移植性挑战。在本文中,我们介绍了TunedCnC,这是一个新颖的,声明性的和灵活的CnC调整框架,用于通过指定层次关系组和分布函数来控制数据和计算的时空放置。 TunedCnC强调关注点分离:领域专家通过定义数据和控件依赖性来指定并行应用程序,而调优专家则指定应如何在给定体系结构上执行应用程序-定义何时何地放置数据和计算。当针对不同的平台或针对不同的性能目标进行调整时,该应用程序保持不变。我们评估了TunedCnC在多个应用程序上的效用,并证明了更改调优规范可能会对应用程序的性能产生重大影响。我们使用并行集合(CnC)声明式并行编程模型的实现来执行我们的评估,但是我们的结果也应该适用于其他数据流任务并行编程模型的调整。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号