首页> 外文会议>IEEE Data Science Workshop >Gradient Coding with Clustering and Multi-Message Communication
【24h】

Gradient Coding with Clustering and Multi-Message Communication

机译:聚类和多消息通信的梯度编码

获取原文

摘要

Gradient descent (GD) methods are commonly employed in machine learning problems to optimize the parameters of the model in an iterative fashion. For problems with massive datasets, computations are distributed to many parallel computing servers (i.e., workers) to speed up GD iterations. While distributed computing can increase the computation speed significantly, the per-iteration completion time is limited by the slowest straggling workers. Coded distributed computing can mitigate straggling workers by introducing redundant computations; however, existing coded computing schemes are mainly designed against persistent stragglers, and partial computations at straggling workers are discarded, leading to wasted computational capacity. In this paper, we propose a novel gradient coding (GC) scheme which allows multiple coded computations to be conveyed from each worker to the master per iteration. We numerically show that the proposed GC with multi-message communication (MMC) together with clustering provides significant improvements in the average completion time (of each iteration), with minimal or no increase in the communication load.
机译:梯度下降(GD)方法通常用于机器学习问题中,以迭代方式优化模型的参数。对于海量数据集的问题,将计算分配到许多并行计算服务器(即工作人员)以加快GD迭代的速度。虽然分布式计算可以显着提高计算速度,但是每次迭代的完成时间受到最慢的散乱工人的限制。编码分布式计算可以通过引入冗余计算来减轻散乱的工作人员;然而,现有的编码计算方案主要是针对持久性散乱者而设计的,散乱性工作者的部分计算被丢弃,从而导致计算能力的浪费。在本文中,我们提出了一种新颖的梯度编码(GC)方案,该方案允许每次迭代将多个编码的计算从每个工作人员传递到主机。我们从数值上显示,所提出的具有多消息通信(MMC)和聚类的GC在(每次迭代)平均完成时间方面提供了显着的改进,而通信负载却没有增加或没有增加。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号