首页> 外文期刊>ACM transactions on computer systems >Mitigating Load Imbalance in Distributed Data Serving with Rack-Scale Memory Pooling
【24h】

Mitigating Load Imbalance in Distributed Data Serving with Rack-Scale Memory Pooling

机译:利用机架规模的内存池减轻分布式数据服务中的负载不平衡

获取原文
获取原文并翻译 | 示例
       

摘要

To provide low-latency and high-throughput guarantees, most large key-value stores keep the data in the memory of many servers. Despite the natural parallelism across lookups, the load imbalance, introduced by heavy skew in the popularity distribution of keys, limits performance. To avoid violating tail latency service-level objectives, systems tend to keep server utilization low and organize the data in micro-shards, which provides units of migration and replication for the purpose of load balancing. These techniques reduce the skew but incur additional monitoring, data replication, and consistency maintenance overheads.In this work, we introduce RackOut, a memory pooling technique that leverages the one-sided remote read primitive of emerging rack-scale systems to mitigate load imbalance while respecting service-level objectives. In RackOut, the data are aggregated at rack-scale granularity, with all of the participating servers in the rack jointly servicing all of the rack's micro-shards. We develop a queuing model to evaluate the impact of RackOut at the datacenter scale. In addition, we implement a RackOut proof-of-concept key-value store, evaluate it on two experimental platforms based on RDMA and Scale-Out NUMA, and use these results to validate the model. We devise two distinct approaches to load balancing within a RackOut unit, one based on random selection of nodes-RackOut_static-and another one based on an adaptive load balancing mechanism-RackOut_adaptive. Our results show that RackOut_static increases throughput by up to 6x for RDMA and 8.6x for Scale-Out NUMA compared to a scale-out deployment, while respecting tight tail latency service-level objectives. RackOut_adaptive improves the throughput by 30% for workloads with 20% of writes over RackOut_static.
机译:为了提供低延迟和高吞吐量的保证,大多数大型键值存储将数据保留在许多服务器的内存中。尽管查找之间自然存在并行性,但是由于键的受欢迎程度分布严重偏斜而导致的负载不平衡仍然限制了性能。为了避免违反尾部等待时间服务级别的目标,系统趋向于保持服务器利用率较低,并以微碎片组织数据,这为负载平衡的目的提供了迁移和复制的单元。这些技术减少了偏斜,但会产生额外的监视,数据复制和一致性维护开销。在这项工作中,我们介绍了RackOut,一种内存池化技术,利用新兴的机架级系统的单面远程读取原语来减轻负载不平衡的同时尊重服务水平目标。在RackOut中,数据以机架级粒度聚合,机架中的所有参与服务器共同为机架的所有微碎片提供服务。我们开发了一个排队模型来评估RackOut在数据中心规模上的影响。此外,我们实现了RackOut概念验证键值存储,并在两个基于RDMA和Scale-Out NUMA的实验平台上对其进行了评估,然后使用这些结果来验证模型。我们设计了两种不同的方法来实现RackOut单元内的负载平衡,一种基于节点的随机选择-RackOut_static-另一种基于自适应负载平衡机制-RackOut_adaptive。我们的结果表明,与横向扩展部署相比,RackOut_static与横向扩展部署相比,RDMA的吞吐量提高了6倍,横向扩展NUMA的吞吐量提高了8.6倍,同时遵守严格的尾部延迟服务级别目标。通过在RackOut_static上进行20%的写入,RackOut_adaptive将工作负载的吞吐量提高了30%。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号