【24h】

Dynamic Thread Partition Algorithm Based on Sharing Data on CMP

机译:基于CMP数据共享的动态线程划分算法

获取原文

摘要

At the level of multi-core processors that share the same cache, data sharing among threads which belong to different cores may not enjoy the benifit of non-uniform cache access because it is difficult to determine which core should be set as the local position of data block while each cache block is setting as one of the core''s local block. Studies have found that the cost of long latency access can be reduced by using a proper thread partition/allocation algorithm [5]. However, at present work, researchers pay little attention to thread partitioning algorithms which can reduce the cost of long latency access. In this paper, we present a dynamic thread partitioning algorithm according to data sharing among threads at the level of cache-shared-multicore processers. In our design, the algorithm makes the best effort to minimize shared block accessed by threads of different cores. Compared with the existing work, our new algorithm achieves a performance improvement. We perform experiments on 4 cores and more than 100 threads and the result show that our algorithm can reduce the interaction of threads belonging to different cores between 30% and 50% over the previously known solutions.
机译:在共享同一高速缓存的多核处理器级别,属于不同内核的线程之间的数据共享可能无法享受非均匀高速缓存访​​问的好处,因为很难确定哪个内核应设置为本地内核。数据块,而每个高速缓存块被设置为内核的本地块之一。研究发现,使用适当的线程分区/分配算法可以降低长等待时间访问的成本[5]。但是,在目前的工作中,研究人员很少关注线程划分算法,该算法可以减少长等待时间访问的成本。在本文中,我们根据缓存共享多核处理器级别上线程之间的数据共享,提出了一种动态线程分区算法。在我们的设计中,该算法将尽最大努力使不同内核的线程访问的共享块最小化。与现有工作相比,我们的新算法实现了性能改进。我们在4个内核和100多个线程上进行了实验,结果表明,与以前已知的解决方案相比,我们的算法可以将属于不同内核的线程之间的交互作用降低30%至50%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号