首页> 外文会议>International Conference on Electrical, Electronics, Communication, Computer, and Optimization Techniques >Performance Comparison of Serial and Parallel Multipliers in Massively Parallel Environment
【24h】

Performance Comparison of Serial and Parallel Multipliers in Massively Parallel Environment

机译:串行平行环境中串行乘法器的性能比较

获取原文

摘要

Computational environment of Deep Learning Neural Networks (DLNNs) is considerably different than that of Conventional computer systems. DLNNs require thousands, if not millions of compute cores compared to one or few in conventional systems. Therefore, there is a need to review the performance issue to gain better understanding of how systems behave in such massively parallel architectures. Precision, speed, memory access, bus contention, resource sharing, chip area etc are some of the key issues that need to be studied in the changed context. Low precision multiplication remains one of the commonly used operations in neural computations. This paper draws reader attention to some interesting results in area-speed tradeoffs when applied to massively parallel architectures. A new low precision fixed point representation is discussed. A hardware accelerators and its software components used in the simulation are briefly discussed. Results show that serial multipliers can perform better than parallel multipliers considering the throughput per unit area of Silicon.
机译:深度学习神经网络(DLNNS)的计算环境比传统计算机系统相比不同。如果不是数百万计数核心,则DLNNS需要数千个,而传统系统中的一个或多个。因此,需要审查性能问题,以更好地了解系统在这种大规模并行架构中的行为方式。精度,速度,内存访问,总线争用,资源共享,芯片区域等是在更改的上下文中需要研究的一些关键问题。低精度乘法仍然是神经计算中常用的操作之一。本文在应用于大规模平行架构时,读者对面积速度折衷的一些有趣的结果引起了一些有趣的结果。讨论了新的低精度定点表示。简要讨论用于模拟中使用的硬件加速器及其软件组件。结果表明,考虑到每单位硅的吞吐量,串行乘法器可以比并行乘法器更好。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号