首页> 外文会议>IEEE International Symposium on Information Theory >Communication Efficient and Byzantine Tolerant Distributed Learning
【24h】

Communication Efficient and Byzantine Tolerant Distributed Learning

机译:高效沟通和拜占庭容忍的分布式学习

获取原文

摘要

We develop a communication-efficient distributed learning algorithm that is robust against Byzantine worker machines. We propose and analyze a distributed gradient-descent algorithm that performs a simple thresholding based on gradient norms to mitigate Byzantine failures. We show the (statistical) error-rate of our algorithm matches that of Yin et al., 2018, which uses more complicated schemes (like coordinate-wise median or trimmed mean). Furthermore, for communication efficiency, we consider a generic class of δ-approximate compressors from Karimireddy et al., 2019, that encompasses signbased compressors and top-k sparsification. Our algorithm uses compressed gradients and gradient norms for aggregation and Byzantine removal respectively. We establish the statistical error rate of the algorithm for arbitrary (convex or non-convex) smooth loss function. We show that, in certain regime of δ, the rate of convergence is not affected by the compression operation. We have experimentally validated our results and shown good performance in convergence for convex (least-square regression) and non-convex (neural network training) problems.
机译:我们开发了一种沟通高效的分布式学习算法,对拜占庭工人机器具有强大。我们提出并分析了一种基于梯度规范来执行简单的阈值算法的分布式梯度 - 下降算法,以减轻拜占庭的故障。我们展示了算法的(统计)误差率与Yin等人的算法相匹配,这是使用更多复杂方案的yin等人的差值(如坐标 - 明智的中位数或修剪平均值)。此外,对于通信效率,我们考虑了来自Karimireddy等人的一般类别的δ - 近似压缩机,2019年,包括符号的压缩机和顶-K稀疏化。我们的算法使用压缩渐变和梯度规范分别用于聚合和拜占庭拆除。我们建立了任意(凸或非凸面)平滑损耗函数算法的统计误差率。我们表明,在δ的某些方案中,收敛速率不受压缩操作的影响。我们已经通过实验验证了我们的结果,并在凸(最小二乘回归)和非凸(神经网络训练)问题中表现出良好的性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号