首页> 外文会议>International conference on algorithms and architectures for parallel processing >Rise the Momentum: A Method for Reducing the Training Error on Multiple GPUs
【24h】

Rise the Momentum: A Method for Reducing the Training Error on Multiple GPUs

机译:升起势头:一种降低多个GPU上训练错误的方法

获取原文

摘要

Deep neural network training is a common issue that is receiving increasing attention in recent years and basically performed on Stochastic Gradient Descent or its variants. Distributed training increases training speed significantly but causes precision loss at the mean time. Increasing batchsize can improve training parallelism in distributed training. However, if the batchsize is too large, it will bring difficulty to training process and introduce more training error. In this paper, we consider controlling the total batchsize and lowering batchsize on each GPU by increasing the number of GPUs in distributed training. We train Resnet50 [4] on CIFAR-10 dataset by different optimizers, such as SGD, Adam and NAG. The experimental results show that large batchsize speeds up convergence to some degree. However, if the batch-size of per GPU is too small, training process fails to converge. Large number of GPUs, which means a small batchsize on each GPU declines the training performance in distributed training. We tried several ways to reduce the training error on multiple GPUs. According to our results, increasing momentum is a well-behaved method in distributed training to improve training performance on condition of multiple GPUs of constant large batchsize.
机译:深度神经网络培训是近年来接受不断关注的常见问题,基本上对随机梯度下降或其变体进行了。分布式训练显着提高训练速度,但在平均时刻导致精度损失。增加批量化可以改善分布式训练中的培训并行性。但是,如果批量化太大,它将带来难以培训过程并引入更多培训错误。在本文中,我们考虑通过增加分布式训练中GPU的数量来控制每个GPU的总批量化和降低批量化。我们通过不同的优化器在Cifar-10数据集上培训Reset50 [4],例如SGD,ADAM和NAG。实验结果表明,大量批量速度速度加快了一定程度。但是,如果每个GPU的批量大小太小,培训过程无法收敛。大量GPU,这意味着对每个GPU的小批量化下降了分布式训练中的培训表现。我们尝试了几种方法来减少多个GPU上的训练错误。根据我们的结果,势头的增加是分布式训练中的一种良好的方法,以改善恒大GPUS的训练性能。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号