首页> 外国专利> EFFICIENT PARALLEL TRAINING OF A NETWORK MODEL ON MULTIPLE GRAPHICS PROCESSING UNITS

EFFICIENT PARALLEL TRAINING OF A NETWORK MODEL ON MULTIPLE GRAPHICS PROCESSING UNITS

机译:多图形处理单元上的网络模型的有效并行训练

摘要

A system and method provides efficient parallel training of a neural network model on multiple graphics processing units. A training module reduces the time and communication overhead of gradient accumulation and parameter updating of the network model in a neural network by overlapping processes in an advantageous way. In a described embodiment, a training module overlaps backpropagation, gradient transfer and accumulation in a Synchronous Stochastic Gradient Decent algorithm on a convolution neural network. The training module collects gradients of multiple layers during backpropagation of training from a plurality of graphics processing units (GPUs), accumulates the gradients on at least one processor and then delivers the gradients of the layers to the plurality of GPUs during the backpropagation of the training. The whole model parameters can then be updated on the GPUs after receipt of the gradient of the last layer.
机译:一种系统和方法在多个图形处理单元上提供了神经网络模型的有效并行训练。训练模块通过重叠过程以有利的方式减少了神经网络中网络模型的梯度累积和参数更新的时间和通信开销。在所描述的实施例中,训练模块在卷积神经网络上的同步随机梯度体面算法中重叠了反向传播,梯度传递和累积。训练模块在训练的反向传播期间从多个图形处理单元(GPU)收集多层的梯度,将梯度累积在至少一个处理器上,然后在训练的反向传播期间将层的梯度传递给多个GPU 。接收到最后一层的梯度后,即可在GPU上更新整个模型参数。

著录项

相似文献

  • 专利
  • 外文文献
  • 中文文献
获取专利

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号