首页> 外文会议>International conference on artificial neural networks >NoSync: Particle Swarm Inspired Distributed DNN Training
【24h】

NoSync: Particle Swarm Inspired Distributed DNN Training

机译:NoSync:粒子群启发式分布式DNN培训

获取原文

摘要

Training deep neural networks on big datasets remains a computational challenge. It can take hundreds of hours to perform and requires distributed computing systems to accelerate. Common distributed data-parallel approaches share a single model across multiple workers, train on different batches, aggregate gradients, and redistribute the new model. In this work, we propose NoSync, a particle swarm optimization inspired alternative where each worker trains a separate model, and applies pressure forcing models to converge. NoSync explores a greater portion of the parameter space and provides resilience to over-fitting. It consistently offers higher accuracy compared to single workers, offers a linear speedup for smaller clusters, and is orthogonal to existing data-parallel approaches.
机译:在大型数据集上训练深度神经网络仍然是计算上的挑战。它可能需要数百小时才能执行,并且需要分布式计算系统才能加速。常见的分布式数据并行方法在多个工作人员之间共享一个模型,对不同批次进行训练,汇总梯度,然后重新分配新模型。在这项工作中,我们提出了NoSync,这是一种受粒子群优化启发的替代方案,其中,每个工人训练一个单独的模型,并应用压力强制模型进行收敛。 NoSync会探索参数空间的较大部分,并为过度拟合提供弹性。与单个工作人员相比,它始终提供更高的精度,为较小的集群提供线性加速,并且与现有的数据并行方法正交。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号