首页> 外文会议>IEEE symposium on parallel and distributed processing >Efficient mapping of randomly sparse neural networks on parallel vector supercomputers
【24h】

Efficient mapping of randomly sparse neural networks on parallel vector supercomputers

机译:在并行矢量超计算机上随机稀疏神经网络的高效映射

获取原文

摘要

This paper presents efficient mappings of large sparse neural networks on a distributed-memory MIMD multicomputer with high performance vector units. We develop parallel vector code for an idealized network and analyze its performance. Our algorithms combine high performance with a reasonable memory requirement. Due to the high cost of scatter/gather operations, generating high performance parallel vector code requires careful attention to details of the representation. We show that vectorization can nevertheless more than quadruple the performance on our modeled supercomputer. Pushing several patterns at a time through the network (batch mode) exposes an extra degree of parallelism which allows us to improve the performance by an additional factor of 4. Vectorization and batch updating therefore yield an order of magnitude performance improvement.
机译:本文介绍了具有高性能向量单元的分布式内存MIMD多电脑的大稀疏神经网络的有效映射。我们为理想化的网络开发并行矢量代码并分析其性能。我们的算法与合理的内存要求相结合了高性能。由于散射/采集操作的高成本,生成高性能并行矢量代码需要仔细关注表示表示的细节。我们表明,VecsiveIzation仍然可以是我们在模型超级计算机上的性能方面的比例。通过网络(批量模式)一次按下几个模式暴露额外的并行度,这使我们能够通过额外的因子来提高性能4.矢量化和批量更新,因此产生了一个数量级性能改进的顺序。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号