首页> 外文期刊>IEEE transactions on very large scale integration (VLSI) systems >NeuronLink: An Efficient Chip-to-Chip Interconnect for Large-Scale Neural Network Accelerators
【24h】

NeuronLink: An Efficient Chip-to-Chip Interconnect for Large-Scale Neural Network Accelerators

机译:NeuronLink:用于大型神经网络加速器的高效芯片互连

获取原文
获取原文并翻译 | 示例

摘要

Large-scale neural network (NN) accelerators typically consist of several processing nodes, which could be implemented as a multi- or many-core chip and organized via a network-on-chip (NoC) to handle the heavy neuron-to-neuron traffic. Multiple NoC-based NN chips are connected through chip-to-chip interconnection networks to further boost the overall neural acceleration capability. Huge amounts of multicast-based traffic travel on-chip or cross chips, making the interconnection network design more challenging and become the bottleneck of the NN system performance and energy. In this article, we propose coupling intrachip and interchip communication techniques, called NeuronLink, for NN accelerators. Regarding the intrachip communication, we propose scoring crossbar arbitration, arbitration interception, and route computation parallelization techniques for virtual-channel routing, leading to a high-throughput NoC with a lower hardware cost for multicast-based traffic. Regarding the interchip communication, we propose a lightweight and NoC-aware chip-to-chip interconnection scheme, enabling efficient interconnection for NoC-based NN chips. In addition, we evaluate the proposed techniques on a four connected NoC-based deep neural network (DNN) chips with four field-programmable gate arrays (FPGAs). The experimental results show that the proposed interconnection network can efficiently manage the data traffic inside DNNs with high-throughput and low-overhead against state-of-the-art interconnects.
机译:大规模神经网络(NN)加速器通常由若干处理节点组成,可以实现为多或多核芯片,并通过片内(NOC)组织以处理重型神经元到神经元交通。基于NN的NN芯片通过芯片到芯片互连网络连接,以进一步提高整体神经加速能力。大量组播基于组播的流量旅行片或交叉芯片,使互连网络设计更具挑战性,成为NN系统性能和能量的瓶颈。在本文中,我们提出了NN加速器的耦合耦合intrachip和Interchip通信技术,称为Neuronlink。关于intachip通信,我们提出了用于虚拟通道路由的横杆仲裁,仲裁拦截和路由计算并行化技术,导致具有较低的基于组播业务的硬件成本的高吞吐量NOC。关于Interchip通信,我们提出了一种轻量级和NOC感知的芯片到芯片互连方案,使基于NOC的NN芯片能够有效地互连。此外,我们评估了具有四个现场可编程门阵列(FPGA)的四个连接的基于NOC的深神经网络(DNN)芯片上的所提出的技术。实验结果表明,所提出的互连网络可以有效地管理DNN内的数据流量,以高吞吐量和用于最先进的互连的低开销。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号