首页> 外文期刊>Computer architecture news >History-Based Arbitration for Fairness in Processor-Interconnect of NUMA Servers
【24h】

History-Based Arbitration for Fairness in Processor-Interconnect of NUMA Servers

机译:基于历史的仲裁,以实现NUMA服务器处理器互连中的公平性

获取原文
获取原文并翻译 | 示例

摘要

NUMA (non-uniform memory access) servers are commonly used in high-performance computing and datacen-ters. Within each server, a processor-interconnect (e.g., Intel QPI, AMD HyperTransport) is used to communicate between the different sockets or nodes. In this work, we explore the impact of the processor-interconnect on overall performance - in particular, the performance unfairness caused by processor-interconnect arbitration. It is well known that locally-fair arbitration does not guarantee globally-fair bandwidth sharing as closer nodes receive more bandwidth in a multi-hop network. However, this work demonstrates that the opposite can occur in a commodity NUMA server where remote nodes receive higher bandwidth (and perform better). We analyze this problem and identify that this occurs because of external concentration used in router micro-architectures for processor-interconnects without globally-aware arbitration. While accessing remote memory can occur in any NUMA system, performance unfairness (or performance variation) is more critical in cloud computing and virtual machines with shared resources. We demonstrate how this unfairness creates significant performance variation when a workload is executed on the Xen virtualization platform. We then provide analysis using synthetic workloads to better understand the source of unfairness and eliminate the impact of other shared resources, including the shared last-level cache and main memory. To provide fairness, we propose a novel, history-based arbitration that tracks the history of arbitration grants made in the previous history window. A weighted arbitration is done based on the history to provide global fairness. Through simulations, we show our proposed history-based arbitration can provide global fairness and minimize the processor-interconnect performance unfairness at low cost.
机译:NUMA(非统一内存访问)服务器通常用于高性能计算和数据中心。在每个服务器中,处理器互连(例如,Intel QPI,AMD HyperTransport)用于在不同的套接字或节点之间进行通信。在这项工作中,我们探讨了处理器互连对整体性能的影响,特别是处理器互连仲裁所导致的性能不公平。众所周知,本地公平仲裁不能保证全球公平的带宽共享,因为更近的节点在多跳网络中会收到更多的带宽。但是,这项工作表明,在商用NUMA服务器中,相反的情况可能会发生,其中远程节点接收到更高的带宽(并且性能更好)。我们分析了此问题,并确定发生这种情况是由于路由器微体系结构中用于处理器互连的外部集中而没有全局意识的仲裁。尽管在任何NUMA系统中都可能发生访问远程内存的问题,但在具有共享资源的云计算和虚拟机中,性能不公平(或性能差异)更为严重。我们演示了在Xen虚拟化平台上执行工作负载时,这种不公平如何导致性能显着变化。然后,我们使用综合工作负载进行分析,以更好地了解不公平的根源,并消除其他共享资源(包括共享的最后一级缓存和主内存)的影响。为了公平起见,我们提出了一种新颖的,基于历史的仲裁,该仲裁可以跟踪在以前的历史窗口中做出的仲裁授权的历史。根据历史记录进行加权仲裁,以提供全球公平性。通过仿真,我们证明了我们提出的基于历史的仲裁可以提供全局公平性,并以低成本将处理器互连性能不公平性降至最低。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号