首页> 外文会议>IEEE International Symposium on High Performance Computer Architecture >HyPar: Towards Hybrid Parallelism for Deep Learning Accelerator Array
【24h】

HyPar: Towards Hybrid Parallelism for Deep Learning Accelerator Array

机译:HyPar:面向深度学习加速器阵列的混合并行

获取原文

摘要

With the rise of artificial intelligence in recent years, Deep Neural Networks (DNNs) have been widely used in many domains. To achieve high performance and energy efficiency, hardware acceleration (especially inference) of DNNs is intensively studied both in academia and industry. However, we still face two challenges: large DNN models and datasets, which incur frequent off-chip memory accesses; and the training of DNNs, which is not well-explored in recent accelerator designs. To truly provide high throughput and energy efficient acceleration for the training of deep and large models, we inevitably need to use multiple accelerators to explore the coarse-grain parallelism, compared to the fine-grain parallelism inside a layer considered in most of the existing architectures. It poses the key research question to seek the best organization of computation and dataflow among accelerators.
机译:近年来,随着人工智能的兴起,深度神经网络(DNN)已广泛应用于许多领域。为了实现高性能和高能效,学术界和工业界都对DNN的硬件加速(尤其是推理)进行了深入研究。但是,我们仍然面临两个挑战:大型DNN模型和数据集,这会导致频繁的片外内存访问;以及DNN的训练,这在最近的加速器设计中还没有得到很好的探索。为了真正提供高吞吐率和高能效的加速来训练深度模型和大型模型,与大多数现有体系结构中考虑的层内的细粒度并行性相比,我们不可避免地需要使用多个加速器来探索粗粒度并行性。它提出了关键的研究问题,以寻求加速器之间最佳的计算和数据流组织。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号