首页> 外文会议>International conference on computational science >Compiler Directed Parallelization of Loops in Scale for Shared-Memory Multiprocessors
【24h】

Compiler Directed Parallelization of Loops in Scale for Shared-Memory Multiprocessors

机译:编译器针​​对共享内存多处理器的循环指向并行化

获取原文
获取外文期刊封面目录资料

摘要

Effective utilization of symmetric shared-memory multiprocessors (SMPs) is predicted on the development of efficient parallel code. Unfortunately, efficient parallelism is not always easy for the programmer to identify. Worse, exploiting such parallelism may directly conflict with optimizations affecting per-processor utilization (i.e. loop recording to improve data locality). Here, we present our experience with a loop-level parallel compiler optimization for SMPs proposed by McKinley [6]. The algorithm uses dependence analysis and a simple model of the target machine, to transform nested loops. The goal of the approach is to promote efficient execution of parallel loops by exposing sources of large-grain parallel work while maintaining per-processor locality. We implement the optimization within the Scale compiler framework, and analyze the performance of multiprocessor code produced for three microbenchmarks.
机译:对对称共享内存多处理器(SMPS)的有效利用预测有效的平行代码的开发。不幸的是,程序员识别的高效并行性并不容易。更糟糕的是,利用这种并行性可能与影响每处理器利用率的优化直接冲突(即循环记录以改善数据局部性)。在这里,我们向McKinley提出的SMPS提供了我们的循环级并行编译器优化的体验[6]。该算法使用依赖性分析和目标机器的简单模型,转换嵌套环。该方法的目标是通过在维持每个处理器局部性的同时暴露大粒并联工作来源来促进平行环路的有效执行。我们在规模编译器框架内实现优化,并分析为三个微型发布标记生成的多处理器代码的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号