首页> 外文期刊>Scientific programming >NUMA-Aware Thread Scheduling for Big Data Transfers over Terabits Network Infrastructure
【24h】

NUMA-Aware Thread Scheduling for Big Data Transfers over Terabits Network Infrastructure

机译:NUMA感知线程调度,用于通过Terabits网络基础结构进行大数据传输

获取原文
           

摘要

The evergrowing trend of big data has led scientists to share and transfer the simulation and analytical data across the geodistributed research and computing facilities. However, the existing data transfer frameworks used for data sharing lack the capability to adopt the attributes of the underlying parallel file systems (PFS). LADS (Layout-Aware Data Scheduling) is an end-to-end data transfer tool optimized for terabit network using a layout-aware data scheduling via PFS. However, it does not consider the NUMA (Nonuniform Memory Access) architecture. In this paper, we propose a NUMA-aware thread and resource scheduling for optimized data transfer in terabit network. First, we propose distributed RMA buffers to reduce memory controller contention in CPU sockets and then schedule the threads based on CPU socket and NUMA nodes inside CPU socket to reduce memory access latency. We design and implement the proposed resource and thread scheduling in the existing LADS framework. Experimental results showed from 21.7% to 44% improvement with memory-level optimizations in the LADS framework as compared to the baseline without any optimization.
机译:大数据的不断增长的趋势已导致科学家在地理分布的研究和计算设施之间共享和传输模拟和分析数据。但是,用于数据共享的现有数据传输框架缺乏采用基础并行文件系统(PFS)属性的能力。 LADS(可识别布局的数据调度)是一种端到端数据传输工具,已针对通过PFS进行可识别布局的数据调度的太比特网络进行了优化。但是,它不考虑NUMA(非均匀内存访问)体系结构。在本文中,我们提出了一种NUMA感知线程和资源调度,以优化兆位网络中的数据传输。首先,我们建议使用分布式RMA缓冲区来减少CPU插槽中的内存控制器争用,然后基于CPU插槽和CPU插槽内的NUMA节点调度线程,以减少内存访问延迟。我们在现有的LADS框架中设计和实现建议的资源和线程调度。实验结果表明,与未进行任何优化的基准相比,LADS框架中的内存级别优化使性能提高了21.7%至44%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号