首页> 外文会议>IEEE/ACM International Symposium on Cluster, Cloud and Grid Computing >A Novel Stochastic Gradient Descent Algorithm Based on Grouping over Heterogeneous Cluster Systems for Distributed Deep Learning
【24h】

A Novel Stochastic Gradient Descent Algorithm Based on Grouping over Heterogeneous Cluster Systems for Distributed Deep Learning

机译:一种新型随机梯度下降算法,基于在分布式深度学习的异构集群系统上进行分组

获取原文

摘要

On heterogeneous cluster systems, the convergence performances of neural network models are greatly troubled by the different performances of machines. In this paper, we propose a novel distributed Stochastic Gradient Descent (SGD) algorithm named Grouping-SGD for distributed deep learning, which converges faster than Sync-SGD, Async-SGD, and Stale-SGD. In Grouping-SGD, machines are partitioned into multiple groups, ensuring that machines in the same group have similar performances. Machines in the same group update the models synchronously, while different groups update the models asynchronously. To improve the performance of Grouping-SGD further, the parameter servers are arranged from fast to slow, and they are responsible for updating the model parameters from the lower layer to the higher layer respectively. The experimental results indicate that Grouping-SGD can achieve 1.2~3.7 times speedups using popular image classification benchmarks: MNIST, Cifar10, Cifar100, and ImageNet, compared to Sync-SGD, Async-SGD, and Stale-SGD.
机译:在异构集群系统上,通过机器的不同性能大大困扰神经网络模型的收敛性能。在本文中,我们提出了一种名为Direct-SGD的分布式随机梯度下降(SGD)算法,用于分布式深度学习,其比Sync-SGD,Async-SGD和Stale-SGD更快地收敛。在分组-SGD中,机器被划分为多个组,确保同一组中的机器具有相似的性能。同一组中的机器同步更新模型,而不同的组则异步更新模型。为了进一步提高分组-SGD的性能,参数服务器从快速排列到慢,并且它们负责分别将从下层从下层更新到更高层的模型参数。实验结果表明,与Sync-SGD,ASYNC-SGD和STALE-SGD相比,分组-SGD可以实现1.2〜3.7倍的加速:MNIST,CIFAR10,CIFAR100和ImageNet。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号