首页> 外文会议>IEEE/IFIP International Conference on Dependable Systems and Networks Workshops >Byzantine Fault-Tolerant Distributed Machine Learning with Norm-Based Comparative Gradient Elimination
【24h】

Byzantine Fault-Tolerant Distributed Machine Learning with Norm-Based Comparative Gradient Elimination

机译:拜占庭式容错分布式机器学习,基于规范的比较梯度消除

获取原文

摘要

This paper considers the Byzantine fault-tolerance problem in distributed stochastic gradient descent (D-SGD) method – a popular algorithm for distributed multi-agent machine learning. In this problem, each agent samples data points independently from a certain data-generating distribution. In the fault-free case, the D-SGD method allows all the agents to learn a mathematical model best fitting the data collectively sampled by all agents. We consider the case when a fraction of agents may be Byzantine faulty. Such faulty agents may not follow a prescribed algorithm correctly, and may render traditional D-SGD method ineffective by sharing arbitrary incorrect stochastic gradients. We propose a norm-based gradient-filter, named comparative gradient elimination (CGE), that robustifies the D-SGD method against Byzantine agents. We show that the CGE gradient-filter guarantees fault-tolerance against a bounded fraction of Byzantine agents under standard stochastic assumptions, and is computationally simpler compared to many existing gradient-filters such as multi-KRUM, geometric median-of-means, and the spectral filters. We empirically show, by simulating distributed learning on neural networks, that the fault-tolerance of CGE is comparable to that of existing gradient-filters. We also empirically show that exponential averaging of stochastic gradients improves the fault-tolerance of a generic gradient-filter.
机译:本文考虑了分布式随机梯度下降(D-SGD)方法的拜占庭容错问题 - 一种分布式多智能体机学习的流行算法。在该问题中,每个代理从某个数据生成分布独立地样本数据点。在无故障情况下,D-SGD方法允许所有代理学习最佳拟合所有代理的数据的数学模型。我们考虑当百分之一的药物可能是拜占庭的错误时。这种错误的代理可能无法正确地遵循规定的算法,并且可以通过共享任意不正确的随机梯度来使传统的D-SGD方法无效。我们提出了一种基于规范的梯度过滤器,命名为比较梯度消除(CGE),其强调了对拜占庭代理的D-SGD方法。我们表明CGE梯度过滤器在标准随机假设下对拜占庭试剂的界限分数保证容错,并且与许多现有梯度过滤器(如多Krum,几何中间均值)和均值相比,计算地是更简单的光谱滤波器。通过模拟神经网络的分布式学习,我们经验显示,CGE的容错与现有梯度过滤器的容错相当。我们还经验证明,随机梯度的指数平均提高了通用梯度过滤器的容错。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号