首页> 外文期刊>Computer communication review >Speedinq up Distributed Request-Response Workflows
【24h】

Speedinq up Distributed Request-Response Workflows

机译:加快分布式请求-响应工作流

获取原文
获取原文并翻译 | 示例
获取外文期刊封面目录资料

摘要

We found that interactive services at Bing have highly variable datacenter-side processing latencies because their processing consists of many sequential stages, paral-lelization across 10s-1000s of servers and aggregation of responses across the network. To improve the tail latency of such services, we use a few building blocks: reissuing laggards elsewhere in the cluster, new policies to return incomplete results and speeding up laggards by giving them more resources. Combining these building blocks to reduce the overall latency is non-trivial because for the same amount of resource (e.g., number of reissues), different stages improve their latency by different amounts. We present Kwiken, a framework that takes an end-to-end view of latency improvements and costs. It decomposes the problem of minimizing latency over a general processing DAG into a manageable optimization over individual stages. Through simulations with production traces, we show sizable gains; the 99th percentile of latency improves by over 50% when just 0.1% of the responses are allowed to have partial results and by over 40% for 25% of the services when just 5% extra resources are used for reissues.
机译:我们发现Bing的交互式服务具有高度可变的数据中心端处理延迟,因为它们的处理包括许多顺序的阶段,跨10s-1000s服务器的并行化以及跨网络的响应聚合。为了改善此类服务的尾部等待时间,我们使用了一些构建块:重新发出集群中其他地方的落后者,返回不完整结果的新策略以及通过为落后者提供更多资源来加快落后者的速度。组合这些构件以减少总体等待时间并非易事,因为对于相同数量的资源(例如,重发次数),不同的阶段可以将等待时间提高不同的数量。我们介绍了Kwiken,这是一个对延迟改进和成本进行端到端查看的框架。它将最小化一般处理DAG上的延迟的问题分解为各个阶段上可管理的优化。通过对生产轨迹的模拟,我们显示出可观的收益;当仅允许0.1%的响应产生部分结果时,延迟的第99个百分位数将提高50%以上;而仅将5%的额外资源用于重新发布时,对于25%的服务,延迟将提高50%以上。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号