【24h】

Using Skip Graphs for Increased NUMA Locality

机译:使用跳过图提高NUMA局部性

获取原文

摘要

High-performance simulations and parallel frameworks often rely on highly scalable, concurrent data structures for system scalability. With an increased availability of NUMA architectures, we present a technique to promote NUMA-aware data parallelism inside a concurrent data structure, bringing significant quantitative and qualitative improvements on NUMA locality, as well as reduced contention for synchronized memory accesses. Our architecture is based on a data-partitioned, concurrent skip graph indexed by thread-local sequential maps. We implemented maps and relaxed priority queues using such technique. Maps show up to 6x higher CAS locality, up to a 68.6% reduction on the number of remote CAS operations, and an increase from 88.3% to 99% on the CAS success rate compared to a control implementation (subject to the same optimizations, and implementation practices). Remote memory accesses are not only reduced in number, but the larger the NUMA distance between threads, the larger the reduction is. Relaxed priority queues implemented using our technique show similar scalability improvements, with provable reduction in contention and decrease in relaxation in one of our implementations.
机译:高性能仿真和并行框架通常依赖于高度可伸缩的并发数据结构来实现系统可伸缩性。随着NUMA体系结构可用性的提高,我们提出了一种在并发数据结构内提升NUMA感知数据并行性的技术,从而对NUMA局部性进行了重大的定量和定性改进,并减少了对同步内存访问的争用。我们的体系结构基于由线程局部顺序图索引的数据分区的并发跳过图。我们使用这种技术实现了地图并放宽了优先级队列。与对照实施相比,地图显示的CAS本地性最高提高了6倍,远程CAS操作的数量减少了68.6%,CAS成功率从88.3%增加到了99%(要进行相同的优化,并且实施实践)。不仅减少了远程内存访问的数量,而且线程之间的NUMA距离越大,减少的幅度也越大。使用我们的技术实现的放宽优先级队列显示出类似的可伸缩性改进,在我们的一种实现中,可证明减少了争用并减少了放宽。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号