首页> 外文会议>IEEE International Conference on Cluster Computing >Cutting the Tail: Designing High Performance Message Brokers to Reduce Tail Latencies in Stream Processing
【24h】

Cutting the Tail: Designing High Performance Message Brokers to Reduce Tail Latencies in Stream Processing

机译:减少尾巴:设计高性能消息代理以减少流处理中的尾巴延迟

获取原文

摘要

Over the last decade, organizations have become heavily reliant on providing near-instantaneous insights to the end user based on vast amounts of data collected from various sources in real-time. In order to accomplish this task, a stream processing pipeline is constructed, which in its most basic form, consists of a Stream Processing Engine (SPE) and a Message Broker (MB). The SPE is responsible for performing actual computations on the data and providing insights from it. MB, on the other hand, acts as an intermediate queue to which data is written by ephemeral sources and then fetched by the SPE to perform computations on. Due to the inherent real-time nature of such a pipeline, low latency is a highly desirable feature for them. Thus, several existing research works in the community focus on improving latency and throughput of the streaming pipeline. However, there is a dearth of studies optimizing the tail latencies of such pipelines. Moreover, the root cause of this high tail latency is still vague. In this paper, we propose a model-based approach to analyze in-depth the reasons behind high tail latency in streaming systems such as Apache Kafka. Having found the MB to be a major contributor of messages with high tail latencies in a streaming pipeline, we design and implement an RDMA-enhanced high-performance MB, called Frieda, with the higher goal of accelerating any arbitrary stream processing pipeline regardless of the SPE used. Our experiments show a reduction of up to 98% in 99.9th percentile latency for microbenchmarks and up to 31% for full-fledged stream processing pipeline constructed using Yahoo! Streaming Benchmark.
机译:在过去的十年中,组织已经非常依赖于根据从各种来源实时收集的大量数据为最终用户提供近乎即时的洞察力。为了完成此任务,构建了一个流处理管道,该流处理管道的最基本形式由流处理引擎(SPE)和消息代理(MB)组成。 SPE负责对数据进行实际计算并提供数据见解。另一方面,MB充当中间队列,临时资源将数据写入该中间队列,然后由SPE提取数据以执行计算。由于此类管道具有固有的实时性,因此低延迟是它们非常需要的功能。因此,社区中的一些现有研究工作集中在改善流传输管道的延迟和吞吐量上。然而,缺乏对这种管道的尾部延迟进行优化的研究。而且,这种高尾部等待时间的根本原因仍然不清楚。在本文中,我们提出了一种基于模型的方法来深入分析诸如Apache Kafka之类的流系统中高尾部延迟的原因。发现MB是流管道中具有高尾延迟的消息的主要贡献者之后,我们设计并实现了RDMA增强的高性能MB,称为Frieda,其更高的目标是加速任何流处理管道,无论使用了SPE。我们的实验显示,微基准测试在99.9%的百分位延迟中最多可减少98%,对于使用Yahoo!构建的完整流处理管道,最多可减少31%。流基准测试。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号