首页> 外文期刊>Journal of Parallel and Distributed Computing >An efficient delay-optimal distributed termination detection algorithm
【24h】

An efficient delay-optimal distributed termination detection algorithm

机译:一种高效的延迟最优分布式终端检测算法

获取原文
获取原文并翻译 | 示例

摘要

Distributed termination detection is a fundamental problem in parallel and distributed computing and numerous schemes with different performance characteristics have been proposed. These schemes, while being efficient with regard to one performance metric, prove to be inefficient in terms of other metrics. A significant drawback shared by all previous methods is that, on most popular topologies, they take Ω(P) time to detect and signal termination after its actual occurrence, where P is the total number of processing elements. Detection delay is arguably the most important metric to optimize, since it is directly related to the amount of idling of computing resources and to the delay in the utilization of results of the underlying computation. In this paper, we present a novel termination detection algorithm that is simultaneously optimal or near-optimal with respect to all relevant performance measures on any topology. In particular, our algorithm has a best-case detection delay of Θ(1) and a finite optimal worst-case detection delay on any topology equal in order terms to the time for an optimal one-to-all broadcast on that topology (which we accurately characterize for an arbitrary topology). On k-ary n-cube tori and meshes, the worst-case delay is Θ(D), where D is the diameter of the target topology. Further, our algorithm has message and computational complexities of Θ(MD + P) in the worst case and, for most applications, Θ(M + P) in the average case—the same as other message-efficient algorithms, and an optimal space complexity of Θ(P), where M is the total number of messages used by the underlying computation. We also give a scheme using counters that greatly reduces the constant associated with the average message and computational complexities, but does not suffer from the counter-overflow problems of other schemes. Finally, unlike some previous schemes, our algorithm does not rely on first-in first-out (FIFO) ordering for message communication to work correctly.
机译:分布式终止检测是并行和分布式计算中的一个基本问题,已经提出了许多具有不同性能特征的方案。这些方案虽然相对于一个性能指标是有效的,但在其他指标方面却证明是无效的。所有先前方法的共同缺点是,在大多数流行的拓扑上,它们花费Ω(P)时间来检测并在实际发生信号终止后终止信号,其中P是处理元素的总数。检测延迟可以说是最重要的优化指标,因为它直接关系到计算资源的闲置量以及底层计算结果的利用延迟。在本文中,我们提出了一种新颖的终止检测算法,该算法对于任何拓扑上的所有相关性能指标而言,同时是最佳或接近最佳的。特别地,我们的算法在任何拓扑上的最佳情况检测延迟为Θ(1),并且在与该拓扑上最佳一对多广播的最佳时间相同的条件下,该拓扑在任何拓扑上均具有有限的最佳最坏情况检测延迟(我们可以准确地表征任意拓扑)。在k元n立方体花托和网格上,最坏情况下的延迟是Θ(D),其中D是目标拓扑的直径。此外,我们的算法在最坏情况下的消息和计算复杂度为Θ(MD + P),在大多数应用中,在一般情况下的消息和计算复杂度为Θ(M + P),与其他消息有效算法相同,并且具有最佳空间复杂度Θ(P),其中M是基础计算使用的消息总数。我们还提供了一种使用计数器的方案,该方案可以大大减少与平均消息和计算复杂性相关的常数,但不会遭受其他方案的计数器溢出问题。最后,与某些先前的方案不同,我们的算法不依赖于先进先出(FIFO)排序来使消息通信正常工作。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号