首页> 外文会议>Annual conference on Neural Information Processing Systems >'Short-Dot': Computing Large Linear Transforms Distributedly Using Coded Short Dot Products
【24h】

'Short-Dot': Computing Large Linear Transforms Distributedly Using Coded Short Dot Products

机译:“短点”:使用编码的短点乘积分布式计算大型线性变换

获取原文

摘要

Faced with saturation of Moore's law and increasing size and dimension of data, system designers have increasingly resorted to parallel and distributed computing to reduce computation time of machine-learning algorithms. However, distributed computing is often bottle necked by a small fraction of slow processors called "stragglers" that reduce the speed of computation because the fusion node has to wait for all processors to complete their processing. To combat the effect of stragglers, recent literature proposes introducing redundancy in computations across processors, e.g., using repetition-based strategies or erasure codes. The fusion node can exploit this redundancy by completing the computation using outputs from only a subset of the processors, ignoring the stragglers. In this paper, we propose a novel technique - that we call "Short-Dot" - to introduce redundant computations in a coding theory inspired fashion, for computing linear transforms of long vectors. Instead of computing long dot products as required in the original linear transform, we construct a larger number of redundant and short dot products that can be computed more efficiently at individual processors. Further, only a subset of these short dot products are required at the fusion node to finish the computation successfully. We demonstrate through probabilistic analysis as well as experiments on computing clusters that Short-Dot offers significant speed-up compared to existing techniques. We also derive trade-offs between the length of the dot-products and the resilience to stragglers (number of processors required to finish), for any such strategy and compare it to that achieved by our strategy.
机译:面对摩尔定律的饱和以及数据大小和维度的增长,系统设计人员越来越多地采用并行和分布式计算来减少机器学习算法的计算时间。但是,分布式计算通常受一小部分称为“散乱者”的慢速处理器的束缚,它们会降低计算速度,因为融合节点必须等待所有处理器完成其处理。为了对抗散乱者的影响,最近的文献提出在处理器之间的计算中引入冗余,例如,使用基于重复的策略或擦除码。融合节点可以通过仅使用一部分处理器的输出来完成计算而忽略散乱者,从而利用这种冗余。在本文中,我们提出了一种新颖的技术-我们称为“短点”-以编码理论启发的方式引入冗余计算,用于计算长向量的线性变换。与构建原始线性变换中所需的长点积不同,我们构造了大量的冗余和短点积,这些积和短点积可以在各个处理器上更有效地进行计算。此外,在融合节点上仅需要这些短点积的子集即可成功完成计算。我们通过概率分析以及在计算集群上的实验证明,与现有技术相比,Short-Dot可以显着提高速度。对于任何这样的策略,我们还得出点积产品的长度和对散乱者的弹性(完成加工所需的处理器数量)之间的折衷,并将其与我们的策略所实现的进行比较。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号