首页> 外文会议>IEEE Conference on Computer Communications Workshops >Poster Abstract: Model Average-based Distributed Training for Sparse Deep Neural Networks
【24h】

Poster Abstract: Model Average-based Distributed Training for Sparse Deep Neural Networks

机译:海报摘要:稀疏深度神经网络的基于模型平均的分布式训练

获取原文
获取外文期刊封面目录资料

摘要

Distributed training of large-scale deep neural net-works(DNNs) is a challenging work for it's time costing and complicated communication. Existing works have achieved scalable performance on GPU clusters for dense DNNs in the computer vision area. However, little progress has been made on the distributed training of sparse DNNs which is commonly used in the area of natural language processing (NLP). In this poster, we introduce SA-HMA, a sparsity-aware hybrid training method for sparse deep models. SA-HMA combines Model Average (MA) and synchronous optimization methods together, expecting to reduce the communication cost for spare model training. The experimental results show that SA-HMA achieves 1.33× speedup over the state-of-the-art work.
机译:大型深度神经网络(DNN)的分布式培训是一项艰巨的工作,因为它耗时且通信复杂。现有工作已在GPU群集上实现了可扩展的性能,可用于计算机视觉领域中的密集DNN。但是,在稀疏DNN的分布式训练方面几乎没有进展,这在自然语言处理(NLP)领域中很普遍。在此海报中,我们介绍了SA-HMA,这是一种用于稀疏深度模型的稀疏感知混合训练方法。 SA-HMA将平均模型(MA)和同步优化方法结合在一起,期望减少备用模型训练的通信成本。实验结果表明,与最新技术相比,SA-HMA的速度提高了1.33倍。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号