...
首页> 外文期刊>Computer architecture news >Optimizing Virtual Machine Consolidation Performance on NUMA Server Architecture for Cloud Workloads
【24h】

Optimizing Virtual Machine Consolidation Performance on NUMA Server Architecture for Cloud Workloads

机译:针对云工作负载在NUMA服务器架构上优化虚拟机整合性能

获取原文
获取原文并翻译 | 示例
           

摘要

Server virtualization and workload consolidation enable multiple workloads to share a single physical server, resulting in significant energy savings and utilization improvements. The shift of physical server architectures to NUMA and the increasing popularity of scale-out cloud applications undermine workload consolidation efficiency and result in overall system degradation. In this work, we characterize the consolidation of cloud workloads on NUMA virtu-alized systems, estimate four different sources of architecture overhead, and explore optimization opportunities beyond the default NUMA-aware hypervisor memory management. Motivated by the observed architectural impact on cloud workload consolidation performance, we propose three optimization techniques incorporating NUMA access overhead into the hypervisor's virtual machine memory allocation and page fault handling routines. Among these, estimation of the memory zone access overhead serves as a foundation for the other two techniques: a NUMA overhead aware buddy allocator and a P2M swap FIFO. Cache hit rate, cycle loss due to cache miss, and IPC serve as indicators to estimate the access cost of each memory node. Our optimized buddy allocator dynamically selects low-overhead memory zones and "proportionally" distributes memory pages across target nodes. The P2M swap FIFO records recently unused <PFN, MFN> lists for mapping exchanges to rebalance memory access pressure within one domain. Our real system based evaluations show a 41.1% performance improvement when consolidating 16-VMs on a 4-socket server (the proposed allocator contributes 22.8% of the performance gain and the P2M swap FIFO accounts for the rest). Furthermore, our techniques can cooperate well with other methods (i.e. vCPU migration) and scale well when varying VM memory size and the number of sockets in a physical host.
机译:服务器虚拟化和工作负载合并使多个工作负载可以共享一个物理服务器,从而显着节省了能源并提高了利用率。物理服务器体系结构向NUMA的转移以及横向扩展云应用程序的日益普及破坏了工作负载整合效率,并导致整体系统性能下降。在这项工作中,我们将表征NUMA虚拟化系统上云工作负载的整合,估算四种不同的体系结构开销来源,并探索除默认的NUMA感知虚拟机管理程序内存管理之外的优化机会。基于观察到的架构对云工作负载合并性能的影响,我们提出了三种优化技术,这些技术将NUMA访问开销纳入了虚拟机管理程序的虚拟机内存分配和页面错误处理例程。其中,对存储区访问开销的估算是其他两种技术的基础:NUMA开销感知伙伴分配器和P2M交换FIFO。高速缓存命中率,由于高速缓存未命中而导致的周期损失以及IPC可以作为估算每个内存节点访问成本的指标。我们优化的伙伴分配器可动态选择低开销的内存区域,并“按比例”在目标节点之间分配内存页面。 P2M交换FIFO记录最近未使用的<PFN,MFN>列表,用于映射交换以重新平衡一个域内的存储器访问压力。我们基于实际系统的评估显示,在4插槽服务器上整合16个VM时,性能提高了41.1%(建议的分配器贡献了22.8%的性能提升,其余部分由P2M交换FIFO承担)。此外,我们的技术可以与其他方法(即vCPU迁移)很好地配合,并且在更改VM内存大小和物理主机中的套接字数量时可以很好地扩展。

著录项

  • 来源
    《Computer architecture news》 |2014年第3期|325-336|共12页
  • 作者

    Ming Liu; Tao Li;

  • 作者单位

    Intelligent Design of Efficient Architectures Laboratory (IDEAL) Department of Electrical and Computer Engineering, University of Florida;

    Intelligent Design of Efficient Architectures Laboratory (IDEAL) Department of Electrical and Computer Engineering, University of Florida;

  • 收录信息
  • 原文格式 PDF
  • 正文语种 eng
  • 中图分类
  • 关键词

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号